code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
def _lowercase( __a : int ): assert ( isinstance(__a , __a ) and number_of_steps > 0 ), f"""number_of_steps needs to be positive integer, your input {number_of_steps}""" if number_of_steps == 1: return 1 a__ , a__ =1, 1 ...
20
import argparse import json import os import sys import tempfile import unittest from argparse import Namespace from dataclasses import dataclass, field from enum import Enum from pathlib import Path from typing import List, Literal, Optional import yaml from transformers import HfArgumentParser, TrainingArguments...
20
1
from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModel...
20
from typing import List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase: List[Any] = logging.get_logger(__name__) _lowerCAmelCase: Any = { 'huggingface/autoformer-tourism-monthly': 'https://huggingface.co/huggingface...
20
1
import PIL.Image import PIL.ImageOps from packaging import version from PIL import Image if version.parse(version.parse(PIL.__version__).base_version) >= version.parse('9.1.0'): _lowerCAmelCase: Tuple = { 'linear': PIL.Image.Resampling.BILINEAR, 'bilinear': PIL.Image.Resampling.BILINEAR...
20
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format...
20
1
import gc import random import unittest import numpy as np import torch from diffusers import ( DDIMScheduler, KandinskyVaaControlnetPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from d...
20
from importlib import import_module from .logging import get_logger _lowerCAmelCase: str = get_logger(__name__) class lowercase_ : def __init__( self , lowercase_ , lowercase_=None) -> Tuple: a__ =attrs or [] if module is not Non...
20
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applic...
20
import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": _lowerCAmelCase: int = argparse.ArgumentParser() parser.add_argument( '--checkpoint_path', default=None, type=str, required=True, help='Path...
20
1
import json import os import tempfile import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSa...
20
import os import re import sys import traceback import warnings from pathlib import Path from typing import Dict, Optional, Union from uuid import uuida from huggingface_hub import HfFolder, ModelCard, ModelCardData, hf_hub_download, whoami from huggingface_hub.file_download import REGEX_COMMIT_HASH from huggingfac...
20
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase: Optional[int] = logging.get_logger(__name__) class lowercase_ (lowercase__ ): snake_case ='timm_backbone' def __init__( self , lowercase_=None , lower...
20
import argparse import json from collections import OrderedDict from functools import partial from pathlib import Path import timm import torch from huggingface_hub import hf_hub_download from transformers import LevitConfig, LevitForImageClassificationWithTeacher, LevitImageProcessor from transformers.utils impor...
20
1
from string import ascii_lowercase, ascii_uppercase def _lowercase( __a : str ): if not sentence: return "" a__ =dict(zip(__a , __a ) ) return lower_to_upper.get(sentence[0] , sentence[0] ) + sentence[1:] if __name__ == "__ma...
20
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models...
20
1
import argparse import torch from transformers import ( WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForAudioFrameClassification, WavaVecaForSequenceClassification, WavaVecaForXVector, logging, ) logging.set_verbosity_info() _lowerCAmelCase: int = logging.get_logger(...
20
import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require...
20
1
import re from flax.core.frozen_dict import freeze from flax.traverse_util import flatten_dict, unflatten_dict from jax.experimental import PartitionSpec as P # Sentinels _lowerCAmelCase: Any = object() # For specifying empty leaf dict `{}` _lowerCAmelCase: Dict = object() ...
20
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BertTokenizer, BlipImageProcessor, Bli...
20
1
import warnings from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch from ...models import UNetaDModel from ...schedulers import RePaintScheduler from ...utils import PIL_INTERPOLATION, logging, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput ...
20
def _lowercase( __a : list[int] ): a__ =len(__a ) for i in range(__a ): for j in range(i + 1 , __a ): if numbers[j] < numbers[i]: a__ , a__ =numbers[j], numbers[i] return numbers ...
20
1
import unittest from transformers import DebertaVaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor fro...
20
import copy import inspect import unittest from transformers import AutoBackbone from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import require_timm, require_torch, torch_device from transformers.utils.import_utils import is_torch_available from ...test_backbone_common...
20
1
from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): import torch if is_tf_available(): i...
20
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) _lowerCAmelCase: Optional[Any] = { 'configuration_swiftformer': [ 'SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'SwiftFormerConfig', 'S...
20
1
def _lowercase( __a : list[int] ): a__ =len(__a ) for i in range(__a ): for j in range(i + 1 , __a ): if numbers[j] < numbers[i]: a__ , a__ =numbers[j], numbers[i] return numbers ...
20
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCAmelCase: str = logging.get_logger(__name__) _lowerCAmelCase: Any = { 'google/bigbird-roberta...
20
1
from copy import deepcopy class lowercase_ : def __init__( self , lowercase_ = None , lowercase_ = None) -> None: if arr is None and size is not None: a__ =size a__ =[0] * size elif arr is not None: ...
20
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_...
20
1
import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transformers.testing_utils import DUMMY...
20
from __future__ import annotations from typing import Any class lowercase_ : def __init__( self , lowercase_) -> None: a__ =num_of_nodes a__ =[] a__ ={} def __UpperCamelCase ( self , lowercase_ , low...
20
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _lowerCAmelCase: Dict = { 'configuration_squeezebert': [ 'SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'SqueezeBertConfig', 'Sq...
20
import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice...
20
1
def _lowercase( __a : Optional[Any] ): if not head: return True # split the list to two parts a__ , a__ =head.next, head while fast and fast.next: a__ =fast.next.next a__ =slow.next a__ =slow.nex...
20
from __future__ import annotations _lowerCAmelCase: str = '#' class lowercase_ : def __init__( self) -> None: a__ ={} def __UpperCamelCase ( self , lowercase_) -> None: a__ =self._trie for char in tex...
20
1
from PIL import Image def _lowercase( __a : Image , __a : int ): a__ =(259 * (level + 255)) / (255 * (259 - level)) def contrast(__a : int ) -> int: return int(128 + factor * (c - 128) ) return img.point(__a ) if _...
20
_lowerCAmelCase: List[str] = 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' def _lowercase( ): a__ =input('Enter message: ' ) a__ =input('Enter key [alphanumeric]: ' ) a__ =input('Encrypt/Decrypt [e/d]: ' ) if mode.lower().startswith('e' ): ...
20
1
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _lowerCAmelCase: Optional[int] = logging.get_logger(__name__) class lowercase_ (lowercase__ , ...
20
from .constants import ( MODEL_NAME, OPTIMIZER_NAME, RNG_STATE_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SCALER_NAME, SCHEDULER_NAME, TORCH_LAUNCH_PARAMS, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .dataclasses import ( BnbQuantizationConfig, ComputeEnvironment, ...
20
1
from __future__ import annotations import math def _lowercase( __a : int ): if num <= 0: a__ =f"""{num}: Invalid input, please enter a positive integer.""" raise ValueError(__a ) a__ =[True] * (num + 1) a__ =[] a__ ...
20
import unittest import numpy as np import torch from torch import nn from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import KandinskyVaaPriorPipeline, PriorTransforme...
20
1
import json import os import re import sys import urllib.request import requests from bsa import BeautifulSoup _lowerCAmelCase: Any = { 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36' ' (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582' } ...
20
from manim import * class lowercase_ (lowercase__ ): def __UpperCamelCase ( self) -> List[Any]: a__ =Rectangle(height=0.5 , width=0.5) a__ =Rectangle(height=0.46 , width=0.46).set_stroke(width=0) a__ =[mem.copy() for...
20
1
import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _lowerCAmelCase: Tuple = '...
20
import argparse import json import os import sys import tempfile import unittest from argparse import Namespace from dataclasses import dataclass, field from enum import Enum from pathlib import Path from typing import List, Literal, Optional import yaml from transformers import HfArgumentParser, TrainingArguments...
20
1
import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, Di...
20
from typing import List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase: List[Any] = logging.get_logger(__name__) _lowerCAmelCase: Any = { 'huggingface/autoformer-tourism-monthly': 'https://huggingface.co/huggingface...
20
1
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models...
20
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format...
20
1
import json import os import unittest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lo...
20
from importlib import import_module from .logging import get_logger _lowerCAmelCase: str = get_logger(__name__) class lowercase_ : def __init__( self , lowercase_ , lowercase_=None) -> Tuple: a__ =attrs or [] if module is not Non...
20
1
import requests _lowerCAmelCase: Union[str, Any] = '' # <-- Put your OpenWeatherMap appid here! _lowerCAmelCase: Union[str, Any] = 'https://api.openweathermap.org/data/2.5/' def _lowercase( __a : str = "Chicago" , __a : str = APPID ): ...
20
import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": _lowerCAmelCase: int = argparse.ArgumentParser() parser.add_argument( '--checkpoint_path', default=None, type=str, required=True, help='Path...
20
1
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class lowercase_ (lowercase__ ): snake_c...
20
import os import re import sys import traceback import warnings from pathlib import Path from typing import Dict, Optional, Union from uuid import uuida from huggingface_hub import HfFolder, ModelCard, ModelCardData, hf_hub_download, whoami from huggingface_hub.file_download import REGEX_COMMIT_HASH from huggingfac...
20
1
from __future__ import annotations from collections.abc import Callable from typing import Generic, TypeVar _lowerCAmelCase: Optional[Any] = TypeVar('T') _lowerCAmelCase: Dict = TypeVar('U') class lowercase_ (Generic[T, U] ): def __init__( self , ...
20
import argparse import json from collections import OrderedDict from functools import partial from pathlib import Path import timm import torch from huggingface_hub import hf_hub_download from transformers import LevitConfig, LevitForImageClassificationWithTeacher, LevitImageProcessor from transformers.utils impor...
20
1
import unittest import numpy as np import requests from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available():...
20
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models...
20
1
import inspect import unittest import numpy as np from transformers import BeitConfig from transformers.testing_utils import require_flax, require_vision, slow from transformers.utils import cached_property, is_flax_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_...
20
import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require...
20
1
import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class lowercase_ (lowercase__ ...
20
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BertTokenizer, BlipImageProcessor, Bli...
20
1
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by a...
20
def _lowercase( __a : list[int] ): a__ =len(__a ) for i in range(__a ): for j in range(i + 1 , __a ): if numbers[j] < numbers[i]: a__ , a__ =numbers[j], numbers[i] return numbers ...
20
1
import torch from diffusers import UnCLIPScheduler from .test_schedulers import SchedulerCommonTest class lowercase_ (lowercase__ ): snake_case =(UnCLIPScheduler,) def __UpperCamelCase ( self , **lowercase_) -> List[str]: a__ ={ 'n...
20
import copy import inspect import unittest from transformers import AutoBackbone from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import require_timm, require_torch, torch_device from transformers.utils.import_utils import is_torch_available from ...test_backbone_common...
20
1
from manim import * class lowercase_ (lowercase__ ): def __UpperCamelCase ( self) -> List[Any]: a__ =Rectangle(height=0.5 , width=0.5) a__ =Rectangle(height=0.46 , width=0.46).set_stroke(width=0) a__ =[mem.copy() for...
20
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) _lowerCAmelCase: Optional[Any] = { 'configuration_swiftformer': [ 'SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'SwiftFormerConfig', 'S...
20
1
import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class lowercase_ (unittest.TestCase ): @property ...
20
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCAmelCase: str = logging.get_logger(__name__) _lowerCAmelCase: Any = { 'google/bigbird-roberta...
20
1
from __future__ import annotations def _lowercase( __a : list[int] , __a : int ): if len(__a ) < k or k < 0: raise ValueError('Invalid Input' ) a__ =a__ =sum(array[:k] ) for i in range(len(__a ) - k ): ...
20
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_...
20
1
from collections import defaultdict from math import ceil, sqrt def _lowercase( __a : int = 100_0000 , __a : int = 10 ): a__ =defaultdict(__a ) for outer_width in range(3 , (t_limit // 4) + 2 ): if outer_width * outer_width > t_l...
20
from __future__ import annotations from typing import Any class lowercase_ : def __init__( self , lowercase_) -> None: a__ =num_of_nodes a__ =[] a__ ={} def __UpperCamelCase ( self , lowercase_ , low...
20
1
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_...
20
import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice...
20
1
import importlib import shutil import threading import warnings from typing import List import fsspec import fsspec.asyn from . import compression from .hffilesystem import HfFileSystem _lowerCAmelCase: List[Any] = importlib.util.find_spec('s3fs') is not None if _has_safs: from .safilesystem i...
20
from __future__ import annotations _lowerCAmelCase: str = '#' class lowercase_ : def __init__( self) -> None: a__ ={} def __UpperCamelCase ( self , lowercase_) -> None: a__ =self._trie for char in tex...
20
1
from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class lowercase_ (lowercase__ ): def __init__( self , lowercase_ , low...
20
_lowerCAmelCase: List[str] = 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' def _lowercase( ): a__ =input('Enter message: ' ) a__ =input('Enter key [alphanumeric]: ' ) a__ =input('Encrypt/Decrypt [e/d]: ' ) if mode.lower().startswith('e' ): ...
20
1
import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow _lowerCAmelCase: List[str] = logging.getLogger() @unittest.skip('Temporarily disable the d...
20
from .constants import ( MODEL_NAME, OPTIMIZER_NAME, RNG_STATE_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SCALER_NAME, SCHEDULER_NAME, TORCH_LAUNCH_PARAMS, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .dataclasses import ( BnbQuantizationConfig, ComputeEnvironment, ...
20
1
from abc import ABC, abstractmethod from typing import List, Optional class lowercase_ (lowercase__ ): def __init__( self) -> List[str]: # test for the above condition self.test() def __UpperCamelCase ( self) -> List[str]: a__ =...
20
import unittest import numpy as np import torch from torch import nn from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import KandinskyVaaPriorPipeline, PriorTransforme...
20
1
import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class lowercase_ (lowercase__ ): snake_case =(PNDMScheduler,) snake_case =(('num_inference_steps', 50),) def __UpperCamelCase ( self , **low...
20
from manim import * class lowercase_ (lowercase__ ): def __UpperCamelCase ( self) -> List[Any]: a__ =Rectangle(height=0.5 , width=0.5) a__ =Rectangle(height=0.46 , width=0.46).set_stroke(width=0) a__ =[mem.copy() for...
20
1
from math import factorial def _lowercase( __a : int = 20 ): a__ =2 * n # middle entry of odd rows starting at row 3 is the solution for n = 1, # 2, 3,... a__ =n // 2 return int(factorial(__a ) / (factorial(__a ) * factorial(n - k )) ) ...
20
import argparse import json import os import sys import tempfile import unittest from argparse import Namespace from dataclasses import dataclass, field from enum import Enum from pathlib import Path from typing import List, Literal, Optional import yaml from transformers import HfArgumentParser, TrainingArguments...
20
1
import argparse import requests import torch from PIL import Image from transformers import ViTMAEConfig, ViTMAEForPreTraining, ViTMAEImageProcessor def _lowercase( __a : List[Any] ): if "cls_token" in name: a__ =name.replace('cls_token' , 'vit.embeddings.cl...
20
from typing import List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase: List[Any] = logging.get_logger(__name__) _lowerCAmelCase: Any = { 'huggingface/autoformer-tourism-monthly': 'https://huggingface.co/huggingface...
20
1
import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFAutoModel, is_tensorflow_text_available, is_tf_available from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.testing_utils import require_tensorflow_text, requir...
20
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format...
20
1
from typing import Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto impor...
20
from importlib import import_module from .logging import get_logger _lowerCAmelCase: str = get_logger(__name__) class lowercase_ : def __init__( self , lowercase_ , lowercase_=None) -> Tuple: a__ =attrs or [] if module is not Non...
20
1
import logging import os from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union from filelock import FileLock from transformers import PreTrainedTokenizer, is_tf_available, is_torch_available _lowerCAmelCase: Optional[Any] = logging.getLogger(__name__) ...
20
import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": _lowerCAmelCase: int = argparse.ArgumentParser() parser.add_argument( '--checkpoint_path', default=None, type=str, required=True, help='Path...
20
1
from __future__ import annotations from typing import Any class lowercase_ : def __init__( self , lowercase_) -> None: a__ =num_of_nodes a__ =[] a__ ={} def __UpperCamelCase ( self , lowercase_ , low...
20
import os import re import sys import traceback import warnings from pathlib import Path from typing import Dict, Optional, Union from uuid import uuida from huggingface_hub import HfFolder, ModelCard, ModelCardData, hf_hub_download, whoami from huggingface_hub.file_download import REGEX_COMMIT_HASH from huggingfac...
20
1
from ..utils import DummyObject, requires_backends class lowercase_ (metaclass=lowercase__ ): snake_case =['flax'] def __init__( self , *lowercase_ , **lowercase_) -> Optional[Any]: requires_backends(self , ['flax']) @classmethod def ...
20
import argparse import json from collections import OrderedDict from functools import partial from pathlib import Path import timm import torch from huggingface_hub import hf_hub_download from transformers import LevitConfig, LevitForImageClassificationWithTeacher, LevitImageProcessor from transformers.utils impor...
20
1
from importlib import import_module from .logging import get_logger _lowerCAmelCase: str = get_logger(__name__) class lowercase_ : def __init__( self , lowercase_ , lowercase_=None) -> Tuple: a__ =attrs or [] if module is not Non...
20
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models...
20
1
from typing import Dict, List, Optional, Union import numpy as np from transformers.utils import is_vision_available from transformers.utils.generic import TensorType from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_re...
20
import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require...
20
1
# tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkou...
20
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BertTokenizer, BlipImageProcessor, Bli...
20
1
from __future__ import annotations _lowerCAmelCase: Union[str, Any] = { 'A': ['B', 'C', 'E'], 'B': ['A', 'D', 'E'], 'C': ['A', 'F', 'G'], 'D': ['B'], 'E': ['A', 'B', 'D'], 'F': ['C'], 'G': ['C'], } class lowercase_ : def __init__( self , ...
20
def _lowercase( __a : list[int] ): a__ =len(__a ) for i in range(__a ): for j in range(i + 1 , __a ): if numbers[j] < numbers[i]: a__ , a__ =numbers[j], numbers[i] return numbers ...
20
1
import math def _lowercase( __a : int ): a__ =[True] * n a__ =False a__ =False a__ =True for i in range(3 , int(n**0.5 + 1 ) , 2 ): a__ =i * 2 while index < n: a...
20
import copy import inspect import unittest from transformers import AutoBackbone from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import require_timm, require_torch, torch_device from transformers.utils.import_utils import is_torch_available from ...test_backbone_common...
20
1
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING _lowerCAmelCase: Tuple = logging.get_logger(__name__) _lowerCAmelCase: Optional[Any] = { 'ut/deta': 'https://huggingface.co/ut/deta/resolve/main/con...
20
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) _lowerCAmelCase: Optional[Any] = { 'configuration_swiftformer': [ 'SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'SwiftFormerConfig', 'S...
20
1
import argparse import json import os import torch from torch import nn from transformers import NllbMoeConfig, NllbMoeModel from transformers.modeling_utils import dtype_byte_size from transformers.utils import WEIGHTS_INDEX_NAME, WEIGHTS_NAME def _lowercase( __a : List[Any] ): a__ ...
20
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCAmelCase: str = logging.get_logger(__name__) _lowerCAmelCase: Any = { 'google/bigbird-roberta...
20
1
import argparse import json import os import sys import tempfile import unittest from argparse import Namespace from dataclasses import dataclass, field from enum import Enum from pathlib import Path from typing import List, Literal, Optional import yaml from transformers import HfArgumentParser, TrainingArguments...
20
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_...
20
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _lowerCAmelCase: List[str] = { 'configuration_roberta': ['ROBERTA_PRETRAINED_CONFIG_A...
20
from __future__ import annotations from typing import Any class lowercase_ : def __init__( self , lowercase_) -> None: a__ =num_of_nodes a__ =[] a__ ={} def __UpperCamelCase ( self , lowercase_ , low...
20
1
def _lowercase( __a : int ): a__ =len(__a ) a__ =sum(__a ) a__ =[[False for x in range(s + 1 )] for y in range(n + 1 )] for i in range(1 , n + 1 ): a__ =True for i in range(1 , s + 1 ): ...
20
import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice...
20
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCAmelCase: Optional[int] = logging.get_logger(__name__) _lowerCAmelCase: Any ...
20
from __future__ import annotations _lowerCAmelCase: str = '#' class lowercase_ : def __init__( self) -> None: a__ ={} def __UpperCamelCase ( self , lowercase_) -> None: a__ =self._trie for char in tex...
20
1
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_t...
20
_lowerCAmelCase: List[str] = 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' def _lowercase( ): a__ =input('Enter message: ' ) a__ =input('Enter key [alphanumeric]: ' ) a__ =input('Encrypt/Decrypt [e/d]: ' ) if mode.lower().startswith('e' ): ...
20
1
class lowercase_ : def __init__( self , lowercase_ , lowercase_=None , lowercase_=None) -> Tuple: a__ =data a__ =previous a__ =next_node def __str__( self) -> str: return F"""{self.data}""" de...
20
from .constants import ( MODEL_NAME, OPTIMIZER_NAME, RNG_STATE_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SCALER_NAME, SCHEDULER_NAME, TORCH_LAUNCH_PARAMS, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .dataclasses import ( BnbQuantizationConfig, ComputeEnvironment, ...
20
1
import argparse import json import numpy import torch from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def _lowercase( __a : str , __a : int ): # Load...
20
import unittest import numpy as np import torch from torch import nn from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import KandinskyVaaPriorPipeline, PriorTransforme...
20
1
def _lowercase( __a : int ): if not isinstance(__a , __a ): a__ =f"""Input value of [number={number}] must be an integer""" raise TypeError(__a ) if number < 0: return False a__ =number * number while num...
20
from manim import * class lowercase_ (lowercase__ ): def __UpperCamelCase ( self) -> List[Any]: a__ =Rectangle(height=0.5 , width=0.5) a__ =Rectangle(height=0.46 , width=0.46).set_stroke(width=0) a__ =[mem.copy() for...
20
1
_lowerCAmelCase: List[str] = 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' def _lowercase( ): a__ =input('Enter message: ' ) a__ =input('Enter key [alphanumeric]: ' ) a__ =input('Encrypt/Decrypt [e/d]: ' ) if mode.lower().startswith('e' ): ...
20
import argparse import json import os import sys import tempfile import unittest from argparse import Namespace from dataclasses import dataclass, field from enum import Enum from pathlib import Path from typing import List, Literal, Optional import yaml from transformers import HfArgumentParser, TrainingArguments...
20
1
_lowerCAmelCase: Union[str, Any] = [4, 1, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] _lowerCAmelCase: int = [3, 7, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] _lowerCAmelCase: Any = { 0: 'Sunday', 1: 'Monday', 2: 'Tuesday', 3: 'Wednesday', 4: 'Thursday', 5: 'Friday', ...
20
from typing import List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase: List[Any] = logging.get_logger(__name__) _lowerCAmelCase: Any = { 'huggingface/autoformer-tourism-monthly': 'https://huggingface.co/huggingface...
20
1
import argparse import json import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import AutoImageProcessor, SwinConfig, SwinForImageClassification def _lowercase( __a : Dict ): a__ =SwinConfig() a__ ...
20
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format...
20
1
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BertTokenizer, BlipImageProcessor, Bli...
20
from importlib import import_module from .logging import get_logger _lowerCAmelCase: str = get_logger(__name__) class lowercase_ : def __init__( self , lowercase_ , lowercase_=None) -> Tuple: a__ =attrs or [] if module is not Non...
20
1
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase: List[str] = logging.get_logger(__name__) _lowerCAmelCase: Any = { 'google/pix2struct-textcaps-base': ( 'https://huggingfac...
20
import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": _lowerCAmelCase: int = argparse.ArgumentParser() parser.add_argument( '--checkpoint_path', default=None, type=str, required=True, help='Path...
20
1
import unittest import numpy as np from transformers.file_utils import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import to...
20
import os import re import sys import traceback import warnings from pathlib import Path from typing import Dict, Optional, Union from uuid import uuida from huggingface_hub import HfFolder, ModelCard, ModelCardData, hf_hub_download, whoami from huggingface_hub.file_download import REGEX_COMMIT_HASH from huggingfac...
20
1
from itertools import product def _lowercase( __a : int , __a : int ): a__ =sides_number a__ =max_face_number * dice_number a__ =[0] * (max_total + 1) a__ =1 a__ =range(__a , max_face_number + 1 ) fo...
20
import argparse import json from collections import OrderedDict from functools import partial from pathlib import Path import timm import torch from huggingface_hub import hf_hub_download from transformers import LevitConfig, LevitForImageClassificationWithTeacher, LevitImageProcessor from transformers.utils impor...
20
1
def _lowercase( __a : float , __a : float ): if density <= 0: raise ValueError('Impossible fluid density' ) if bulk_modulus <= 0: raise ValueError('Impossible bulk modulus' ) return (bulk_modulus / density) ** 0.5 if __name...
20
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models...
20
1
import logging import os import threading import time try: import warnings except ImportError: _lowerCAmelCase: Optional[int] = None try: import msvcrt except ImportError: _lowerCAmelCase: List[str] = None try: import fcntl except ImportError: _lowerCAmelCase: Dict ...
20
import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require...
20
1
import unittest import numpy as np from transformers import RobertaConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models...
20
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BertTokenizer, BlipImageProcessor, Bli...
20
1
from __future__ import annotations def _lowercase( __a : list[list[int]] ): # preprocessing the first row for i in range(1 , len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1 , ...
20
def _lowercase( __a : list[int] ): a__ =len(__a ) for i in range(__a ): for j in range(i + 1 , __a ): if numbers[j] < numbers[i]: a__ , a__ =numbers[j], numbers[i] return numbers ...
20
1
import os import re import sys import traceback import warnings from pathlib import Path from typing import Dict, Optional, Union from uuid import uuida from huggingface_hub import HfFolder, ModelCard, ModelCardData, hf_hub_download, whoami from huggingface_hub.file_download import REGEX_COMMIT_HASH from huggingfac...
20
import copy import inspect import unittest from transformers import AutoBackbone from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import require_timm, require_torch, torch_device from transformers.utils.import_utils import is_torch_available from ...test_backbone_common...
20
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCAmelCase: str = logging.get_logger(__name__) _lowerCAmelCase: Optional[Any] = { 'kssteven/ibe...
20
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) _lowerCAmelCase: Optional[Any] = { 'configuration_swiftformer': [ 'SWIFTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'SwiftFormerConfig', 'S...
20
1
import datasets _lowerCAmelCase: List[str] = '\\n@InProceedings{conneau2018xnli,\n author = "Conneau, Alexis\n and Rinott, Ruty\n and Lample, Guillaume\n and Williams, Adina\n and Bowman, Samuel R.\n and Schwenk, Holg...
20
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCAmelCase: str = logging.get_logger(__name__) _lowerCAmelCase: Any = { 'google/bigbird-roberta...
20
1
import gc import unittest from diffusers import FlaxDPMSolverMultistepScheduler, FlaxStableDiffusionPipeline from diffusers.utils import is_flax_available, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate...
20
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_...
20
1
def _lowercase( __a : int = 100_0000 ): a__ =[i - 1 for i in range(limit + 1 )] for i in range(2 , limit + 1 ): if phi[i] == i - 1: for j in range(2 * i , limit + 1 , __a ): phi[j] -=...
20
from __future__ import annotations from typing import Any class lowercase_ : def __init__( self , lowercase_) -> None: a__ =num_of_nodes a__ =[] a__ ={} def __UpperCamelCase ( self , lowercase_ , low...
20
1
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, Dat...
20
import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice...
20
1
def _lowercase( __a : str , __a : str ): assert x is not None assert y is not None a__ =len(__a ) a__ =len(__a ) # declaring the array for storing the dp values a__ =[[0] * (n + 1) for _ in range(m + 1 )] # noqa:...
20
from __future__ import annotations _lowerCAmelCase: str = '#' class lowercase_ : def __init__( self) -> None: a__ ={} def __UpperCamelCase ( self , lowercase_) -> None: a__ =self._trie for char in tex...
20
1
import argparse import json import os import torch from transformers import LukeConfig, LukeModel, LukeTokenizer, RobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def _lowercase( __a : Union[str, Any] , __a : Union[str, Any] , _...
20
_lowerCAmelCase: List[str] = 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' def _lowercase( ): a__ =input('Enter message: ' ) a__ =input('Enter key [alphanumeric]: ' ) a__ =input('Encrypt/Decrypt [e/d]: ' ) if mode.lower().startswith('e' ): ...
20
1
from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...tes...
20
from .constants import ( MODEL_NAME, OPTIMIZER_NAME, RNG_STATE_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SCALER_NAME, SCHEDULER_NAME, TORCH_LAUNCH_PARAMS, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .dataclasses import ( BnbQuantizationConfig, ComputeEnvironment, ...
20
1
from __future__ import annotations from fractions import Fraction from math import gcd, sqrt def _lowercase( __a : int ): a__ =int(number**0.5 ) return number == sq * sq def _lowercase( __a : int , __a : int , __a : int...
20
import unittest import numpy as np import torch from torch import nn from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import KandinskyVaaPriorPipeline, PriorTransforme...
20
1
_lowerCAmelCase: Optional[int] = tuple[float, float, float] _lowerCAmelCase: List[str] = tuple[float, float, float] def _lowercase( __a : Pointad , __a : Pointad ): a__ =end_pointa[0] - end_pointa[0] a__ =end_pointa[1] -...
20
from manim import * class lowercase_ (lowercase__ ): def __UpperCamelCase ( self) -> List[Any]: a__ =Rectangle(height=0.5 , width=0.5) a__ =Rectangle(height=0.46 , width=0.46).set_stroke(width=0) a__ =[mem.copy() for...
20
1
import unittest import numpy as np from diffusers import LMSDiscreteScheduler, OnnxStableDiffusionInpaintPipeline from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineT...
20
import argparse import json import os import sys import tempfile import unittest from argparse import Namespace from dataclasses import dataclass, field from enum import Enum from pathlib import Path from typing import List, Literal, Optional import yaml from transformers import HfArgumentParser, TrainingArguments...
20
1
class lowercase_ : def __init__( self , lowercase_ , lowercase_) -> str: a__ =name a__ =val def __str__( self) -> Tuple: return F"""{self.__class__.__name__}({self.name}, {self.val})""" def __lt__( self , ...
20
from typing import List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase: List[Any] = logging.get_logger(__name__) _lowerCAmelCase: Any = { 'huggingface/autoformer-tourism-monthly': 'https://huggingface.co/huggingface...
20
1
import math import unittest def _lowercase( __a : int ): assert isinstance(__a , __a ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number <...
20
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format...
20
1
from __future__ import annotations import math def _lowercase( __a : float , __a : int ): a__ =u for i in range(1 , __a ): a__ =temp * (u - i) return temp def _lowercase( ): a__ =int(input('enter t...
20
from importlib import import_module from .logging import get_logger _lowerCAmelCase: str = get_logger(__name__) class lowercase_ : def __init__( self , lowercase_ , lowercase_=None) -> Tuple: a__ =attrs or [] if module is not Non...
20
1
from typing import List, Optional, Union import numpy as np import PIL import torch from PIL import Image from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils ...
20
import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": _lowerCAmelCase: int = argparse.ArgumentParser() parser.add_argument( '--checkpoint_path', default=None, type=str, required=True, help='Path...
20
1
import unittest from transformers import is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_avail...
20
import os import re import sys import traceback import warnings from pathlib import Path from typing import Dict, Optional, Union from uuid import uuida from huggingface_hub import HfFolder, ModelCard, ModelCardData, hf_hub_download, whoami from huggingface_hub.file_download import REGEX_COMMIT_HASH from huggingfac...
20
1
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, PNDMScheduler, StableDiffusionLDMaDPipeline, UNetaDConditionModel, ) from diffusers.utils import nightly, slow, t...
20
import argparse import json from collections import OrderedDict from functools import partial from pathlib import Path import timm import torch from huggingface_hub import hf_hub_download from transformers import LevitConfig, LevitForImageClassificationWithTeacher, LevitImageProcessor from transformers.utils impor...
20
1
def _lowercase( __a : list[int] ): if not nums: # Makes sure that the list is not empty raise ValueError('List is empty' ) a__ =sum(__a ) / len(__a ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(__a ) ...
20
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models...
20
1
_lowerCAmelCase: dict[str, float] = { "km/h": 1.0, "m/s": 3.6, "mph": 1.609_344, "knot": 1.852, } _lowerCAmelCase: dict[str, float] = { "km/h": 1.0, "m/s": 0.277_777_778, "mph": 0.621_371_192, "knot": 0.539_956_803, } def _lowercase( __a ...
20
import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require...
20
1
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(lowercase__ ) , 'Tatoeba dir...
20
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BertTokenizer, BlipImageProcessor, Bli...
20
1
import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_token...
20
def _lowercase( __a : list[int] ): a__ =len(__a ) for i in range(__a ): for j in range(i + 1 , __a ): if numbers[j] < numbers[i]: a__ , a__ =numbers[j], numbers[i] return numbers ...
20
1
import math def _lowercase( __a : int ): a__ =0 a__ =0 while num > 0: a__ =num % 8 a__ =octal + (remainder * math.floor(math.pow(10 , __a ) )) counter += 1 a__ =math.floor(num ...
20
import copy import inspect import unittest from transformers import AutoBackbone from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import require_timm, require_torch, torch_device from transformers.utils.import_utils import is_torch_available from ...test_backbone_common...
20
1