code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
import importlib import inspect import os import re # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py _lowerCamelCase ="""src/transformers""" # This is to make sure the transformers module imported is the one in t...
681
import PIL.Image import PIL.ImageOps from packaging import version from PIL import Image if version.parse(version.parse(PIL.__version__).base_version) >= version.parse("""9.1.0"""): _lowerCamelCase ={ """linear""": PIL.Image.Resampling.BILINEAR, """bilinear""": PIL.Image.Resampling.BILI...
681
1
import math def _a ( lowerCamelCase ): return math.sqrt(lowerCamelCase ) * math.sqrt(lowerCamelCase ) == num def _a ( lowerCamelCase ): lowerCamelCase : str = 0 lowerCamelCase : Tuple = n while left <= right: lowerCamelCase ...
681
from typing import Optional from torch import nn from .transformer_ad import TransformeraDModel, TransformeraDModelOutput class A__ ( nn.Module): def __init__( self , __magic_name__ = 1_6 , __magic_name__ = 8_8 , __magic_name__ = None , __magic_name__ = 1 ...
681
1
from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional from packaging import version if TYPE_CHECKING: from ... import PreTrainedTokenizer, TensorType from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...u...
681
import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _lowerCamelCase ="""▁""" _lowerCamelCase...
681
1
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import convert_to_rgb, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI...
681
from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration _lowerCamelCase =HfArgumentParser(InitializationArguments) _lowerCamelCase =parser.parse_args() # Load codeparrot tokenizer trained for Python code tokeni...
681
1
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDCond...
681
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from transformers import TensorFlowBenchmark, TensorFlowBenchmarkArguments @requir...
681
1
def _a ( lowerCamelCase ): lowerCamelCase : Dict = len(lowerCamelCase ) while cur > 1: # Find the maximum number in arr lowerCamelCase : Dict = arr.index(max(arr[0:cur] ) ) # Reverse from 0 to mi lowerCamelCase : Any = arr[m...
681
import unittest from transformers.testing_utils import CaptureStdout from transformers.tools.python_interpreter import evaluate def _a ( lowerCamelCase ): return x + 2 class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): lowerCamelCase : List...
681
1
from collections import defaultdict from pathlib import Path import pandas as pd from rouge_cli import calculate_rouge_path from utils import calculate_rouge _lowerCamelCase =[ """Prosecutor: \"No videos were used in the crash investigation\" German papers say they saw a cell phone video of the""" """ ...
681
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """edbeeching/decision-transformer-gym-hopper-medium""": ( """https://huggingface.co/edbeeching/decision-transformer-gym-hopper-medium/resolve/main/co...
681
1
def _a ( lowerCamelCase ): if number < 0: raise ValueError("""number must not be negative""" ) return number & (number - 1) == 0 if __name__ == "__main__": import doctest doctest.testmod()
681
import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig _lowerCamelCase =logging.get_logger(__name__) class A__ : def __init__( self , __magic_name__ ...
681
1
import cmath import math def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ): lowerCamelCase : Dict = math.radians(lowerCamelCase ) lowerCamelCase : Optional[int] = math.radians(lowerCamelCase ) # Convert voltage...
681
import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def _a ( lowerCamelCase, lowerCamelCase ): lowerCamelCase : List[Any] = F'''{sampling_rate}''' lowerCamelCase : Optional[int] = """1""" low...
681
1
import re import string from collections import Counter import sacrebleu import sacremoses from packaging import version import datasets _lowerCamelCase =""" @inproceedings{xu-etal-2016-optimizing, title = {Optimizing Statistical Machine Translation for Text Simplification}, authors={Xu, Wei and Napole...
681
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingfac...
681
1
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from transformers import TensorFlowBenchmark, TensorFlowBenchmarkArguments @requir...
681
from __future__ import annotations def _a ( lowerCamelCase ): lowerCamelCase : Union[str, Any] = str(lowerCamelCase ) return n == n[::-1] def _a ( lowerCamelCase = 100_0000 ): lowerCamelCase : Any = 0 for i in range(1, lowerCame...
681
1
def _a ( lowerCamelCase ): lowerCamelCase : List[str] = [0] * len(lowerCamelCase ) lowerCamelCase : Optional[Any] = [] lowerCamelCase : Tuple = [1] * len(lowerCamelCase ) for values in graph.values(): for i in values: indegree[...
681
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTImageProcessor, ViTMSNConfig, ViTMSNModel from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD torch.set_grad_enabled(False) def ...
681
1
from .configuration_bert_masked import MaskedBertConfig from .modeling_bert_masked import ( MaskedBertForMultipleChoice, MaskedBertForQuestionAnswering, MaskedBertForSequenceClassification, MaskedBertForTokenClassification, MaskedBertModel, ) from .modules import *
681
def _a ( lowerCamelCase ): if num < 0: return False lowerCamelCase : int = num lowerCamelCase : int = 0 while num > 0: lowerCamelCase : str = rev_num * 10 + (num % 10) num //= 10 return num_copy == rev_num if __name__ == "__main...
681
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """google/vit-base-patch16-224""...
681
from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_torch_available from ...utils import OptionalDependencyNotAvailable _lowerCamelCase ={ """configuration_gpt_neox_japanese""": ["""GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GPTNeoXJapaneseConfig"""], """tokenization_g...
681
1
def _a ( lowerCamelCase, lowerCamelCase ): print("""\nThe shortest path matrix using Floyd Warshall algorithm\n""" ) for i in range(lowerCamelCase ): for j in range(lowerCamelCase ): if dist[i][j] != float("""inf""" ): print(int(dist[i][j] ), end="""\t""" ) else: ...
681
import copy import random from transformers import CLIPTokenizer class A__ ( __SCREAMING_SNAKE_CASE): def __init__( self , *__magic_name__ , **__magic_name__ ): super().__init__(*__magic_name__ , **__magic_name__ ) lowerCamelCase : Dict ...
681
1
import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig _lowerCamelCase =logging.get_logger(__name__) class A__ : def __init__( self , __magic_name__ ...
681
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_availa...
681
1
import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TextClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_co...
681
from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common i...
681
1
def _a ( lowerCamelCase ): # if the collection is empty, returns empty if collection == []: return [] # get some information about the collection lowerCamelCase : Any = len(lowerCamelCase ) lowerCamelCase : str = max(lowerCamelCase ) lowerCamelCase ...
681
import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): # Initialise PyTorch model low...
681
1
import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def _a ( lowerCamelCase, lowerCamelCase ): lowerCamelCase : List[Any] = F'''{sampling_rate}''' lowerCamelCase : Optional[int] = """1""" low...
681
import argparse import requests import torch from PIL import Image from transformers import CLIPProcessor, GroupViTConfig, GroupViTModel def _a ( lowerCamelCase ): # vision encoder if "img_encoder.pos_embed" in name: lowerCamelCase : Tuple = name.replace("""img_encoder.pos_...
681
1
import argparse import fairseq import torch from torch import nn from transformers import ( MBartaaTokenizer, MBartConfig, MBartForCausalLM, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_ver...
681
from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax.numpy as jnp from jax import random from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .scheduling_utils_flax import FlaxSchedulerMixin @flax.struct.dataclass cla...
681
1
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import tensorflow as tf from transformers import AutoTokenizer, TFAutoModelForSeqaSeqLM ...
681
from itertools import product from cva import COLOR_BGR2GRAY, cvtColor, imread, imshow, waitKey from numpy import dot, exp, mgrid, pi, ravel, square, uinta, zeros def _a ( lowerCamelCase, lowerCamelCase ): lowerCamelCase : List[str] = k_size // 2 lowerCamelCase , ...
681
1
import argparse import shutil import time from json import JSONDecodeError from logging import getLogger from pathlib import Path from typing import Dict, List import torch from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from utils import ( ...
681
import pytest _lowerCamelCase ="""__dummy_dataset1__""" _lowerCamelCase =""" import json import os import datasets REPO_URL = \"https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/\" URLS = {\"train\": REPO_URL + \"wikiann-bn-train.jsonl\", \"validation\": REPO_URL + \"wikiann-bn...
681
1
import importlib import shutil import threading import warnings from typing import List import fsspec import fsspec.asyn from . import compression from .hffilesystem import HfFileSystem _lowerCamelCase =importlib.util.find_spec("""s3fs""") is not None if _has_safs: from .safilesystem import SaFileSystem...
681
import PIL.Image import PIL.ImageOps from packaging import version from PIL import Image if version.parse(version.parse(PIL.__version__).base_version) >= version.parse("""9.1.0"""): _lowerCamelCase ={ """linear""": PIL.Image.Resampling.BILINEAR, """bilinear""": PIL.Image.Resampling.BILI...
681
1
def _a ( lowerCamelCase ): lowerCamelCase : List[Any] = n ** (1 / 3) return (val * val * val) == n if __name__ == "__main__": print(perfect_cube(2_7)) print(perfect_cube(4))
681
from typing import Optional from torch import nn from .transformer_ad import TransformeraDModel, TransformeraDModelOutput class A__ ( nn.Module): def __init__( self , __magic_name__ = 1_6 , __magic_name__ = 8_8 , __magic_name__ = None , __magic_name__ = 1 ...
681
1
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_video_inputs if is_torch_available(): import torc...
681
import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _lowerCamelCase ="""▁""" _lowerCamelCase...
681
1
import unittest from transformers.testing_utils import CaptureStdout from transformers.tools.python_interpreter import evaluate def _a ( lowerCamelCase ): return x + 2 class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): lowerCamelCase : List...
681
from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration _lowerCamelCase =HfArgumentParser(InitializationArguments) _lowerCamelCase =parser.parse_args() # Load codeparrot tokenizer trained for Python code tokeni...
681
1
from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration _lowerCamelCase =HfArgumentParser(InitializationArguments) _lowerCamelCase =parser.parse_args() # Load codeparrot tokenizer trained for Python code tokeni...
681
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from transformers import TensorFlowBenchmark, TensorFlowBenchmarkArguments @requir...
681
1
import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class A__ ( tf.keras.layers.Layer): def __init__( self , __magic_name__...
681
import unittest from transformers.testing_utils import CaptureStdout from transformers.tools.python_interpreter import evaluate def _a ( lowerCamelCase ): return x + 2 class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): lowerCamelCase : List...
681
1
from collections import OrderedDict from typing import List, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """google/efficientnet-b7"...
681
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """edbeeching/decision-transformer-gym-hopper-medium""": ( """https://huggingface.co/edbeeching/decision-transformer-gym-hopper-medium/resolve/main/co...
681
1
import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer,...
681
import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig _lowerCamelCase =logging.get_logger(__name__) class A__ : def __init__( self , __magic_name__ ...
681
1
def _a ( lowerCamelCase ): if n == 1 or not isinstance(lowerCamelCase, lowerCamelCase ): return 0 elif n == 2: return 1 else: lowerCamelCase : List[str] = [0, 1] for i in range(2, n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return s...
681
import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def _a ( lowerCamelCase, lowerCamelCase ): lowerCamelCase : List[Any] = F'''{sampling_rate}''' lowerCamelCase : Optional[int] = """1""" low...
681
1
from ..utils import is_flax_available, is_torch_available if is_torch_available(): from .autoencoder_kl import AutoencoderKL from .controlnet import ControlNetModel from .dual_transformer_ad import DualTransformeraDModel from .modeling_utils import ModelMixin from .prior_transformer imp...
681
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingfac...
681
1
import os import random import sys from . import cryptomath_module as cryptoMath # noqa: N812 from . import rabin_miller as rabinMiller # noqa: N812 def _a ( ): print("""Making key files...""" ) make_key_files("""rsa""", 1024 ) print("""Key files generation successful.""" ) def ...
681
from __future__ import annotations def _a ( lowerCamelCase ): lowerCamelCase : Union[str, Any] = str(lowerCamelCase ) return n == n[::-1] def _a ( lowerCamelCase = 100_0000 ): lowerCamelCase : Any = 0 for i in range(1, lowerCame...
681
1
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: fr...
681
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTImageProcessor, ViTMSNConfig, ViTMSNModel from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD torch.set_grad_enabled(False) def ...
681
1
class A__ : def __init__( self , __magic_name__ ): lowerCamelCase : Optional[int] = set_counts lowerCamelCase : Optional[int] = max(__magic_name__ ) lowerCamelCase : int = len(__magic_name__ ) lowerCamelC...
681
def _a ( lowerCamelCase ): if num < 0: return False lowerCamelCase : int = num lowerCamelCase : int = 0 while num > 0: lowerCamelCase : str = rev_num * 10 + (num % 10) num //= 10 return num_copy == rev_num if __name__ == "__main...
681
1
import argparse import requests import torch from PIL import Image from transformers import CLIPProcessor, GroupViTConfig, GroupViTModel def _a ( lowerCamelCase ): # vision encoder if "img_encoder.pos_embed" in name: lowerCamelCase : Tuple = name.replace("""img_encoder.pos_...
681
from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_torch_available from ...utils import OptionalDependencyNotAvailable _lowerCamelCase ={ """configuration_gpt_neox_japanese""": ["""GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GPTNeoXJapaneseConfig"""], """tokenization_g...
681
1
import argparse import json import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import AutoImageProcessor, SwinConfig, SwinForImageClassification def _a ( lowerCamelCase ): lowerCamelCase : Tuple = SwinCo...
681
import copy import random from transformers import CLIPTokenizer class A__ ( __SCREAMING_SNAKE_CASE): def __init__( self , *__magic_name__ , **__magic_name__ ): super().__init__(*__magic_name__ , **__magic_name__ ) lowerCamelCase : Dict ...
681
1
import argparse import logging import os import time import timeit import datasets import numpy as np import pycuda.autoinit # noqa: F401 import pycuda.driver as cuda import tensorrt as trt import torch from absl import logging as absl_logging from accelerate import Accelerator from datasets import load_dataset, loa...
681
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_availa...
681
1
import argparse import re from flax.traverse_util import flatten_dict, unflatten_dict from tax import checkpoints from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model from transformers.uti...
681
from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common i...
681
1
import requests _lowerCamelCase ="""https://newsapi.org/v1/articles?source=bbc-news&sortBy=top&apiKey=""" def _a ( lowerCamelCase ): # fetching a list of articles in json format lowerCamelCase : List[str] = requests.get(_NEWS_API + bbc_news_api_key ).json() # each arti...
681
import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): # Initialise PyTorch model low...
681
1
import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger _lowerCamelCase =get_logger(__name__) class A__ ( enum.Enum): _UpperCAmelCase : Any = """all_checks""" _UpperCAmelCase ...
681
import argparse import requests import torch from PIL import Image from transformers import CLIPProcessor, GroupViTConfig, GroupViTModel def _a ( lowerCamelCase ): # vision encoder if "img_encoder.pos_embed" in name: lowerCamelCase : Tuple = name.replace("""img_encoder.pos_...
681
1
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingfac...
681
from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax.numpy as jnp from jax import random from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .scheduling_utils_flax import FlaxSchedulerMixin @flax.struct.dataclass cla...
681
1
import argparse import os import torch from transformers.utils import WEIGHTS_NAME _lowerCamelCase =["""small""", """medium""", """large"""] _lowerCamelCase ="""lm_head.decoder.weight""" _lowerCamelCase ="""lm_head.weight""" def _a ( lowerCamelCase, lowerCamelCase ): lowerCamel...
681
from itertools import product from cva import COLOR_BGR2GRAY, cvtColor, imread, imshow, waitKey from numpy import dot, exp, mgrid, pi, ravel, square, uinta, zeros def _a ( lowerCamelCase, lowerCamelCase ): lowerCamelCase : List[str] = k_size // 2 lowerCamelCase , ...
681
1
from __future__ import annotations from cmath import sqrt def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): if a == 0: raise ValueError("""Coefficient 'a' must not be zero.""" ) lowerCamelCase : str = b * b - 4 * a * c lowerCamelCase : str ...
681
import pytest _lowerCamelCase ="""__dummy_dataset1__""" _lowerCamelCase =""" import json import os import datasets REPO_URL = \"https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/\" URLS = {\"train\": REPO_URL + \"wikiann-bn-train.jsonl\", \"validation\": REPO_URL + \"wikiann-bn...
681
1
import collections import os import re from pathlib import Path _lowerCamelCase ="""src/transformers""" # Matches is_xxx_available() _lowerCamelCase =re.compile(R"""is\_([a-z_]*)_available()""") # Catches a one-line _import_struct = {xxx} _lowerCamelCase =re.compile(R"""^_import_structure\s+=\s+\{([^\}...
681
import PIL.Image import PIL.ImageOps from packaging import version from PIL import Image if version.parse(version.parse(PIL.__version__).base_version) >= version.parse("""9.1.0"""): _lowerCamelCase ={ """linear""": PIL.Image.Resampling.BILINEAR, """bilinear""": PIL.Image.Resampling.BILI...
681
1
import numpy as np _lowerCamelCase =[ ["""a""", """b""", """c""", """d""", """e"""], ["""f""", """g""", """h""", """i""", """k"""], ["""l""", """m""", """n""", """o""", """p"""], ["""q""", """r""", """s""", """t""", """u"""], ["""v""", """w""", """x""", """y""", """z"""], ] class A__ ...
681
from typing import Optional from torch import nn from .transformer_ad import TransformeraDModel, TransformeraDModelOutput class A__ ( nn.Module): def __init__( self , __magic_name__ = 1_6 , __magic_name__ = 8_8 , __magic_name__ = None , __magic_name__ = 1 ...
681
1
def _a ( lowerCamelCase, lowerCamelCase ): return price * (1 + tax_rate) if __name__ == "__main__": print(f'''{price_plus_tax(1_0_0, 0.25) = }''') print(f'''{price_plus_tax(125.50, 0.05) = }''')
681
import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _lowerCamelCase ="""▁""" _lowerCamelCase...
681
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _lowerCamelCase ={ """configuration_roberta_prelayernorm""": [ """ROBERTA_PRELAYERNORM_PRETRAINED_CONFIG_ARCHIVE_MAP""", ...
681
from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration _lowerCamelCase =HfArgumentParser(InitializationArguments) _lowerCamelCase =parser.parse_args() # Load codeparrot tokenizer trained for Python code tokeni...
681
1
import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import BatchEncoding, MarianTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import is_sentencepiece_available, is_tf_available, is_torch_available ...
681
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from transformers import TensorFlowBenchmark, TensorFlowBenchmarkArguments @requir...
681
1
import warnings from diffusers import StableDiffusionImgaImgPipeline # noqa F401 warnings.warn( """The `image_to_image.py` script is outdated. Please use directly `from diffusers import""" """ StableDiffusionImg2ImgPipeline` instead.""" )
681
import unittest from transformers.testing_utils import CaptureStdout from transformers.tools.python_interpreter import evaluate def _a ( lowerCamelCase ): return x + 2 class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): lowerCamelCase : List...
681
1
def _a ( ): return 1 def _a ( lowerCamelCase ): return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def _a ( lowerCamelCase ): return 0 if x < 0 else five_pence(x - 5 ) + two_pence(lowerCamelCase ) def _a ( lowerCamelCase ): return 0 if x ...
681
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """edbeeching/decision-transformer-gym-hopper-medium""": ( """https://huggingface.co/edbeeching/decision-transformer-gym-hopper-medium/resolve/main/co...
681
1
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokenizer else: _lowerCamelCase...
681
import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig _lowerCamelCase =logging.get_logger(__name__) class A__ : def __init__( self , __magic_name__ ...
681
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _lowerCamelCase ={"""configuration_swin""": ["""SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """SwinConfig""", """SwinOnnxConfig"""]} try: if not is_torch_available(): ...
681
import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def _a ( lowerCamelCase, lowerCamelCase ): lowerCamelCase : List[Any] = F'''{sampling_rate}''' lowerCamelCase : Optional[int] = """1""" low...
681
1
from .testing import ( are_the_same_tensors, execute_subprocess_async, require_bnb, require_cpu, require_cuda, require_huggingface_suite, require_mps, require_multi_gpu, require_multi_xpu, require_safetensors, require_single_gpu, require_single_xpu, require_torch_min...
681
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingfac...
681
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """google/pegasus-large""": """https://huggingface.co/google/pegasus-large/resolve/main/config.json""", # See all PEGASUS models at https://huggingface.co...
681
from __future__ import annotations def _a ( lowerCamelCase ): lowerCamelCase : Union[str, Any] = str(lowerCamelCase ) return n == n[::-1] def _a ( lowerCamelCase = 100_0000 ): lowerCamelCase : Any = 0 for i in range(1, lowerCame...
681
1
import argparse import collections import json import os import re import string import sys import numpy as np _lowerCamelCase =re.compile(R"""\b(a|an|the)\b""", re.UNICODE) _lowerCamelCase =None def _a ( ): lowerCamelCase : int = argparse.ArgumentParser("""Official eva...
681
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTImageProcessor, ViTMSNConfig, ViTMSNModel from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD torch.set_grad_enabled(False) def ...
681
1
from __future__ import annotations def _a ( lowerCamelCase ): # preprocessing the first row for i in range(1, len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1, len(lowerCamelCase ) ): matrix[i][0] += matrix[i - 1][0]...
681
def _a ( lowerCamelCase ): if num < 0: return False lowerCamelCase : int = num lowerCamelCase : int = 0 while num > 0: lowerCamelCase : str = rev_num * 10 + (num % 10) num //= 10 return num_copy == rev_num if __name__ == "__main...
681
1
import logging import os from logging import ( CRITICAL, # NOQA DEBUG, # NOQA ERROR, # NOQA FATAL, # NOQA INFO, # NOQA NOTSET, # NOQA WARN, # NOQA WARNING, # NOQA ) from typing import Optional from tqdm import auto as tqdm_lib _lowerCamelCase ={ """debug""": logging.D...
681
from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_torch_available from ...utils import OptionalDependencyNotAvailable _lowerCamelCase ={ """configuration_gpt_neox_japanese""": ["""GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GPTNeoXJapaneseConfig"""], """tokenization_g...
681
1
import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): lowerCamelCase : Union[str, Any] = [ """safety_checker/pytorch_model.bin""", """safety...
681
import copy import random from transformers import CLIPTokenizer class A__ ( __SCREAMING_SNAKE_CASE): def __init__( self , *__magic_name__ , **__magic_name__ ): super().__init__(*__magic_name__ , **__magic_name__ ) lowerCamelCase : Dict ...
681
1
from random import randint, random def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase = False, lowerCamelCase = False, lowerCamelCase = 5, ): lowerCamelCase : Optional[int] = [[-1] * number_of_cells] # Create a highway without any...
681
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_availa...
681
1
SCREAMING_SNAKE_CASE__ : Tuple = { """a""": """AAAAA""", """b""": """AAAAB""", """c""": """AAABA""", """d""": """AAABB""", """e""": """AABAA""", """f""": """AABAB""", """g""": """AABBA""", """h""": """AABBB""", """i""": """ABAAA""", """j""": """BBBAA""", """...
0
from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common i...
681
0
import unittest import numpy as np import torch from .utils_summarization import build_mask, compute_token_type_ids, process_story, truncate_or_pad class __lowerCamelCase (unittest.TestCase ): def snake_case_ ( self: List[Any] ): '''simp...
1
import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): # Initialise PyTorch model low...
681
0
import multiprocessing from typing import TYPE_CHECKING, Optional, Union from .. import Dataset, Features, config from ..formatting import query_table from ..packaged_modules.sql.sql import Sql from ..utils import logging from .abc import AbstractDatasetInputStream if TYPE_CHECKING: import sqli...
2
import argparse import requests import torch from PIL import Image from transformers import CLIPProcessor, GroupViTConfig, GroupViTModel def _a ( lowerCamelCase ): # vision encoder if "img_encoder.pos_embed" in name: lowerCamelCase : Tuple = name.replace("""img_encoder.pos_...
681
0
'''simple docstring''' from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature fr...
3
from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax.numpy as jnp from jax import random from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .scheduling_utils_flax import FlaxSchedulerMixin @flax.struct.dataclass cla...
681
0
"""simple docstring""" def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : str , _UpperCAmelCase : int ): lowerCAmelCase = word.split() def justify(_UpperCAmelCase : list , _UpperCAmelCase : int , _UpperCAmelCase : int ) -> str: lowerCAmelCase ...
4
from itertools import product from cva import COLOR_BGR2GRAY, cvtColor, imread, imshow, waitKey from numpy import dot, exp, mgrid, pi, ravel, square, uinta, zeros def _a ( lowerCamelCase, lowerCamelCase ): lowerCamelCase : List[str] = k_size // 2 lowerCamelCase , ...
681
0
'''simple docstring''' import random import unittest import numpy as np import torch from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionUpscalePipeline, PNDMScheduler, ) from diffusers.utils...
5
import pytest _lowerCamelCase ="""__dummy_dataset1__""" _lowerCamelCase =""" import json import os import datasets REPO_URL = \"https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/\" URLS = {\"train\": REPO_URL + \"wikiann-bn-train.jsonl\", \"validation\": REPO_URL + \"wikiann-bn...
681
0
import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig _lowerCamelCase = logging.get_logger(__name__) class UpperCamelCase_ : def __init__( self :...
6
import PIL.Image import PIL.ImageOps from packaging import version from PIL import Image if version.parse(version.parse(PIL.__version__).base_version) >= version.parse("""9.1.0"""): _lowerCamelCase ={ """linear""": PIL.Image.Resampling.BILINEAR, """bilinear""": PIL.Image.Resampling.BILI...
681
0
"""simple docstring""" from __future__ import annotations def _snake_case ( _snake_case : list , _snake_case : int ) -> Optional[Any]: '''simple docstring''' if len(_snake_case ) <= 1 or n <= 1: return insert_next(_snake_case ...
7
from typing import Optional from torch import nn from .transformer_ad import TransformeraDModel, TransformeraDModelOutput class A__ ( nn.Module): def __init__( self , __magic_name__ = 1_6 , __magic_name__ = 8_8 , __magic_name__ = None , __magic_name__ = 1 ...
681
0
'''simple docstring''' import inspect import unittest from transformers import BitConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_co...
8
import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _lowerCamelCase ="""▁""" _lowerCamelCase...
681
0
import warnings from .generation import TFGenerationMixin class __lowerCAmelCase ( UpperCAmelCase_ ): """simple docstring""" warnings.warn( "Importing `TFGenerationMixin` from `src/transformers/generation_tf_utils.py` is deprecated and will " "be removed in Transformers v5...
9
from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration _lowerCamelCase =HfArgumentParser(InitializationArguments) _lowerCamelCase =parser.parse_args() # Load codeparrot tokenizer trained for Python code tokeni...
681
0
from __future__ import annotations def _snake_case ( __snake_case , __snake_case = None ): _UpperCamelCase = word_bank or [] # create a table _UpperCamelCase = len(__snake_case ) + 1 _UpperCamelCase = [] for _ in range(__snake_case ): ...
10
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from transformers import TensorFlowBenchmark, TensorFlowBenchmarkArguments @requir...
681
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase_ = { "configuration_nllb_moe": [ "NLLB_MOE_PRETRAINED_CONFIG_ARCHIVE_MAP", "NllbMoeConfig", ] } try: if not is_torch_av...
11
import unittest from transformers.testing_utils import CaptureStdout from transformers.tools.python_interpreter import evaluate def _a ( lowerCamelCase ): return x + 2 class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): lowerCamelCase : List...
681
0
from typing import Dict, List from nltk.translate import gleu_score import datasets from datasets import MetricInfo lowerCamelCase__ : Tuple = """\ @misc{wu2016googles, title={Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation}, autho...
12
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """edbeeching/decision-transformer-gym-hopper-medium""": ( """https://huggingface.co/edbeeching/decision-transformer-gym-hopper-medium/resolve/main/co...
681
0
'''simple docstring''' A__ : str = [ [0, 16, 13, 0, 0, 0], [0, 0, 10, 12, 0, 0], [0, 4, 0, 0, 14, 0], [0, 0, 9, 0, 0, 20], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] def UpperCAmelCase__ ( UpperCAmelCase_ : Dict , UpperCAmelCase_ : A...
13
import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig _lowerCamelCase =logging.get_logger(__name__) class A__ : def __init__( self , __magic_name__ ...
681
0
from __future__ import annotations from fractions import Fraction def __UpperCAmelCase ( __a : int ,__a : int ) -> bool: """simple docstring""" return ( num != den and num % 10 == den // 10 and (num // 10) / (den % 10) == num / den ) ...
14
import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def _a ( lowerCamelCase, lowerCamelCase ): lowerCamelCase : List[Any] = F'''{sampling_rate}''' lowerCamelCase : Optional[int] = """1""" low...
681
0
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer A : Dict = logging.get_logger(__name__) A :...
15
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingfac...
681
0
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging __A : Dict = logging.get_logger(__name__) __A : Optional[Any] ...
16
from __future__ import annotations def _a ( lowerCamelCase ): lowerCamelCase : Union[str, Any] = str(lowerCamelCase ) return n == n[::-1] def _a ( lowerCamelCase = 100_0000 ): lowerCamelCase : Any = 0 for i in range(1, lowerCame...
681
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase_ : Any = { '''configuration_roberta''': ['''ROBERTA_PRETRAINED_CONFIG_ARCH...
17
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTImageProcessor, ViTMSNConfig, ViTMSNModel from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD torch.set_grad_enabled(False) def ...
681
0
'''simple docstring''' import gc import threading import time import psutil import torch class lowerCAmelCase_ : def __init__( self ) -> Union[str, Any]: _lowerCAmelCase = psutil.Process() _lowerCAmelCase = False def _snake_case ( self ) ...
18
def _a ( lowerCamelCase ): if num < 0: return False lowerCamelCase : int = num lowerCamelCase : int = 0 while num > 0: lowerCamelCase : str = rev_num * 10 + (num % 10) num //= 10 return num_copy == rev_num if __name__ == "__main...
681
0
"""simple docstring""" import argparse import torch from transformers import ( EncodecConfig, EncodecFeatureExtractor, EncodecModel, logging, ) # checkpoints downloaded from: # https://dl.fbaipublicfiles.com/encodec/v0/encodec_24khz-d7cc33bc.th # https://huggingface.co/fac...
19
from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_torch_available from ...utils import OptionalDependencyNotAvailable _lowerCamelCase ={ """configuration_gpt_neox_japanese""": ["""GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GPTNeoXJapaneseConfig"""], """tokenization_g...
681
0
from __future__ import annotations import math def _lowercase( __a : int ): if num <= 0: a__ =f"""{num}: Invalid input, please enter a positive integer.""" raise ValueError(__a ) a__ =[True] * (num + 1) a__ =[] a__ ...
20
import copy import random from transformers import CLIPTokenizer class A__ ( __SCREAMING_SNAKE_CASE): def __init__( self , *__magic_name__ , **__magic_name__ ): super().__init__(*__magic_name__ , **__magic_name__ ) lowerCamelCase : Dict ...
681
0
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ : Any = logging.get_logger(__name__) UpperCAmelCase_ : ...
21
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_availa...
681
0
'''simple docstring''' import numpy as np def snake_case_ (UpperCamelCase : Optional[int] , UpperCamelCase : Union[str, Any] , UpperCamelCase : Optional[Any] , UpperCamelCase : List[str] , UpperCamelCase : Dict ): '''simple docst...
22
from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common i...
681
0
import os import time import pytest from datasets.utils.filelock import FileLock, Timeout def _snake_case (__lowercase): UpperCamelCase_ = FileLock(str(tmpdir / 'foo.lock')) UpperCamelCase_ = FileLock(str(tmpdir / 'foo.lock')) UpperCamelCase_ = 0....
23
import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): # Initialise PyTorch model low...
681
0
'''simple docstring''' from __future__ import annotations from math import pi, sqrt def _UpperCamelCase (_lowerCamelCase : float , _lowerCamelCase : float )-> tuple: '''simple docstring''' if inductance <= 0: raise ValueError('''Inductance cannot be...
24
import argparse import requests import torch from PIL import Image from transformers import CLIPProcessor, GroupViTConfig, GroupViTModel def _a ( lowerCamelCase ): # vision encoder if "img_encoder.pos_embed" in name: lowerCamelCase : Tuple = name.replace("""img_encoder.pos_...
681
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a_ = { 'configuration_clap': [ 'CLAP_PRETRAINED_MODEL_ARCHIVE_LIST', 'ClapAudioConfig', 'ClapConfig', 'ClapTextConfig', ], 'processing_clap': ['...
25
from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax.numpy as jnp from jax import random from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .scheduling_utils_flax import FlaxSchedulerMixin @flax.struct.dataclass cla...
681
0
'''simple docstring''' from __future__ import annotations def _a ( _lowerCamelCase ) -> bool: """simple docstring""" __snake_case : Union[str, Any] = len(_lowerCamelCase ) # We need to create solution object to sav...
26
from itertools import product from cva import COLOR_BGR2GRAY, cvtColor, imread, imshow, waitKey from numpy import dot, exp, mgrid, pi, ravel, square, uinta, zeros def _a ( lowerCamelCase, lowerCamelCase ): lowerCamelCase : List[str] = k_size // 2 lowerCamelCase , ...
681
0
import sys from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be c...
27
import pytest _lowerCamelCase ="""__dummy_dataset1__""" _lowerCamelCase =""" import json import os import datasets REPO_URL = \"https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/\" URLS = {\"train\": REPO_URL + \"wikiann-bn-train.jsonl\", \"validation\": REPO_URL + \"wikiann-bn...
681
0
'''simple docstring''' import itertools import math def lowercase__( __UpperCamelCase: int ): """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == ...
28
import PIL.Image import PIL.ImageOps from packaging import version from PIL import Image if version.parse(version.parse(PIL.__version__).base_version) >= version.parse("""9.1.0"""): _lowerCamelCase ={ """linear""": PIL.Image.Resampling.BILINEAR, """bilinear""": PIL.Image.Resampling.BILI...
681
0
"""simple docstring""" import requests from bsa import BeautifulSoup def lowercase ( lowerCAmelCase__ = "AAPL" ): lowerCamelCase_ = f"https://in.finance.yahoo.com/quote/{symbol}?s={symbol}" lowerCamelCase_ = BeautifulSoup(requests.get(lowerCAmelCase__ ).text ,'''html.parser''' ...
29
from typing import Optional from torch import nn from .transformer_ad import TransformeraDModel, TransformeraDModelOutput class A__ ( nn.Module): def __init__( self , __magic_name__ = 1_6 , __magic_name__ = 8_8 , __magic_name__ = None , __magic_name__ = 1 ...
681
0
import unittest from transformers import JukeboxTokenizer from transformers.testing_utils import require_torch class __a( unittest.TestCase ): """simple docstring""" lowerCAmelCase = JukeboxTokenizer lowerCAmelCase = { '''artist''': '''Zac Brown Band''', ...
30
import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _lowerCamelCase ="""▁""" _lowerCamelCase...
681
0
import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_DOCSTRING, BertEmbeddings, BertLay...
31
from arguments import InitializationArguments from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, HfArgumentParser # Configuration _lowerCamelCase =HfArgumentParser(InitializationArguments) _lowerCamelCase =parser.parse_args() # Load codeparrot tokenizer trained for Python code tokeni...
681
0
import argparse import torch from transformers import GPTaConfig, GPTaModel, load_tf_weights_in_gpta from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def A__ ( SCREAMING_SNAKE_CASE_ : List[str] , SCREAMING_SNAKE_CASE_ : Any , SCRE...
32
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from transformers import TensorFlowBenchmark, TensorFlowBenchmarkArguments @requir...
681
0
import itertools import string from collections.abc import Generator, Iterable def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase ) -> Generator[tuple[str, ...], None, None]: snake_case__ = iter(__lowerCAmelCase ) while True: snake_case__ = t...
33
import unittest from transformers.testing_utils import CaptureStdout from transformers.tools.python_interpreter import evaluate def _a ( lowerCamelCase ): return x + 2 class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): lowerCamelCase : List...
681
0
"""simple docstring""" import argparse import json import logging import os import sys from unittest.mock import patch from transformers.testing_utils import TestCasePlus, get_gpu_count, slow SCREAMING_SNAKE_CASE_ = [ os.path.join(os.path.dirname(__file__), dirname) for dirname in [ ...
34
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ """edbeeching/decision-transformer-gym-hopper-medium""": ( """https://huggingface.co/edbeeching/decision-transformer-gym-hopper-medium/resolve/main/co...
681
0
a_ :int = 6_55_21 def a ( A__ ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[Any] = 1 SCREAMING_SNAKE_CASE__ : str = 0 for plain_chr in plain_text: SCREAMING_SNAKE_CASE__ : Any ...
35
import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig _lowerCamelCase =logging.get_logger(__name__) class A__ : def __init__( self , __magic_name__ ...
681
0
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTImageProcessor, ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logg...
36
import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def _a ( lowerCamelCase, lowerCamelCase ): lowerCamelCase : List[Any] = F'''{sampling_rate}''' lowerCamelCase : Optional[int] = """1""" low...
681
0
def UpperCamelCase_ ( __a = 50 ) -> int: a__ : Tuple = [[0] * 3 for _ in range(length + 1 )] for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): ...
37
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingfac...
681
0
'''simple docstring''' from __future__ import annotations def UpperCamelCase__ ( __magic_name__ : str , __magic_name__ : str ) -> bool: '''simple docstring''' snake_case__ : Union[str, Any] = get_failure_array(__magic_name__ ) # 2) Step throu...
38
from __future__ import annotations def _a ( lowerCamelCase ): lowerCamelCase : Union[str, Any] = str(lowerCamelCase ) return n == n[::-1] def _a ( lowerCamelCase = 100_0000 ): lowerCamelCase : Any = 0 for i in range(1, lowerCame...
681
0
from __future__ import annotations from random import random from typing import Generic, TypeVar lowerCAmelCase_ = TypeVar('''KT''') lowerCAmelCase_ = TypeVar('''VT''') class snake_case_ ( Generic[KT, VT] ): '''simple docstring''' de...
39
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTImageProcessor, ViTMSNConfig, ViTMSNModel from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD torch.set_grad_enabled(False) def ...
681
0