code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
import math def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase ): if 0 not in (x, y): # We use the relation x^y = y*log10(x), where 10 is the base. return y * math.logaa(lowerCamelCase ) else: if x == 0: # 0 raised to any number is 0 ...
21
from typing import List from .keymap import KEYMAP, get_character def lowerCAmelCase_ ( lowerCamelCase ): def decorator(lowerCamelCase ): __magic_name__ : str =getattr(lowerCamelCase , """handle_key""" , [] ) handle += [key] s...
21
1
from math import cos, sin, sqrt, tau from audio_filters.iir_filter import IIRFilter def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase = 1 / sqrt(2 ) ): __magic_name__ : Tuple =tau * frequency / samplerate __magic_name__ : ...
21
import os import jsonlines import numpy as np from tqdm import tqdm UpperCAmelCase_ : Dict = 2048 UpperCAmelCase_ : int = 4096 UpperCAmelCase_ : Any = 42 UpperCAmelCase_ : Optional[int] = os.environ.pop("PROCESS_TRAIN", "false") UpperCA...
21
1
import os import jsonlines import numpy as np from tqdm import tqdm UpperCAmelCase_ : Dict = 2048 UpperCAmelCase_ : int = 4096 UpperCAmelCase_ : Any = 42 UpperCAmelCase_ : Optional[int] = os.environ.pop("PROCESS_TRAIN", "false") UpperCA...
21
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ : Tuple = logging.get_logger(__name__) UpperCAmelCase_ : List[str] = { ...
21
1
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ : Union[str, Any] = logging.get_logger(__name__) UpperCAmelCase_ : List[str] = { "tanreinama/GPTSAN-2.8B-spout_is_uniform": ( "https://huggingface.co/tanreinama...
21
from pathlib import Path import fire from tqdm import tqdm def lowerCAmelCase_ ( lowerCamelCase="ro" , lowerCamelCase="en" , lowerCamelCase="wmt16" , lowerCamelCase=None ): try: import datasets except (ModuleNotFoundError, ImportError): raise I...
21
1
import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device ...
21
from __future__ import annotations from fractions import Fraction def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase ): return ( num != den and num % 10 == den // 10 and (num // 10) / (den % 10) == num / den ) def lowerCAmelCase_ ( lowerCamelCase ...
21
1
import math from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass # ...
21
from collections.abc import Callable from math import pi, sqrt from random import uniform from statistics import mean def lowerCAmelCase_ ( lowerCamelCase ): # A local function to see if a dot lands in the circle. def is_in_circle(lowerCamelCase , lowerCamelCase ) -> bool:...
21
1
import inspect import unittest from datasets import load_dataset from packaging import version from transformers import BeitConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from t...
21
import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class __A ( tf.keras.layers.Layer ): def __init__( self ...
21
1
# We ignore warnings about stepping the scheduler since we step it ourselves during gradient accumulation import warnings from .state import AcceleratorState, GradientState warnings.filterwarnings("ignore", category=UserWarning, module="torch.optim.lr_scheduler") class __A : d...
21
import math import tensorflow as tf from packaging import version def lowerCAmelCase_ ( lowerCamelCase ): __magic_name__ : str =tf.convert_to_tensor(lowerCamelCase ) __magic_name__ : List[str] =0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0...
21
1
import secrets from random import shuffle from string import ascii_letters, ascii_lowercase, ascii_uppercase, digits, punctuation def lowerCAmelCase_ ( lowerCamelCase = 8 ): __magic_name__ : Optional[Any] =ascii_letters + digits + punctuation return "".join(secre...
21
from collections.abc import Sequence def lowerCAmelCase_ ( lowerCamelCase = None ): if nums is None or not nums: raise ValueError("""Input sequence should not be empty""" ) __magic_name__ : str =nums[0] for i in range(1 , len(lowerCamelCase ) ...
21
1
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import TransformeraDModel, VQDiffusionPipeline, VQDiffusionScheduler, VQModel from diffusers.pipelines.vq_diffusion.pipeline_vq_diffusion import LearnedClass...
21
from dataclasses import dataclass, field from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import pyarrow as pa if TYPE_CHECKING: from .features import FeatureType @dataclass class __A : UpperCamelCase = 42 UpperCamelCase ...
21
1
import logging import re import pytorch_quantization import pytorch_quantization.nn as quant_nn import torch from pytorch_quantization import calib from pytorch_quantization.tensor_quant import QuantDescriptor UpperCAmelCase_ : Union[str, Any] = logging.getLogger(__name__) Upper...
21
from sklearn.metrics import matthews_corrcoef import datasets UpperCAmelCase_ : Dict = "\nCompute the Matthews correlation coefficient (MCC)\n\nThe Matthews correlation coefficient is used in machine learning as a\nmeasure of the quality of binary and multiclass classifications. It take...
21
1
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ : Any = logging.get_logger(__name__) class __A ( UpperCamelCase__ ): UpperCamelCase = """encoder-decoder""" UpperCamelCase =...
21
import importlib import torch import yaml from omegaconf import OmegaConf from taming.models.vqgan import VQModel def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase=False ): __magic_name__ : Optional[int] =OmegaConf.load(lowerCamelCase ) if display: ...
21
1
from unittest import TestCase from datasets import Dataset from minhash_deduplication import deduplicate_dataset, make_duplicate_clusters def lowerCAmelCase_ ( ): __magic_name__ : List[str] ={ """repo_name""": ["""test_repo1""", """test_repo2""", """test_re...
21
import unittest from accelerate import debug_launcher from accelerate.test_utils import require_cpu, test_ops, test_script @require_cpu class __A ( unittest.TestCase ): def A__ ( self :Tuple ): '''simple docstring''' debug_launcher(test_s...
21
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ : Tuple = logging.get_logger(__name__) UpperCAmelCase_ : List[str] = { ...
21
UpperCAmelCase_ : Tuple = 0 # The first color of the flag. UpperCAmelCase_ : Any = 1 # The second color of the flag. UpperCAmelCase_ : str = 2 # The third color of the flag. UpperCAmelCase_ : Tuple = (red, white, blue) def lowerCAmel...
21
1
import json import os from functools import lru_cache from typing import TYPE_CHECKING, List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational im...
21
# DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from typing import Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import randn_tensor from .scheduling_utils import SchedulerMixin ...
21
1
import unittest import numpy as np import torch from diffusers import DDIMPipeline, DDIMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow, torch_device from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDIT...
21
from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy UpperCAmelCase_ : Dict = logging.get_logger(_...
21
1
from __future__ import annotations from fractions import Fraction def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase ): return ( num != den and num % 10 == den // 10 and (num // 10) / (den % 10) == num / den ) def lowerCAmelCase_ ( lowerCamelCase ...
21
import inspect import unittest import torch import torch.nn as nn from accelerate.hooks import ( AlignDevicesHook, ModelHook, SequentialHook, add_hook_to_module, attach_align_device_hook, remove_hook_from_module, remove_hook_from_submodules, ) from accelerate.test_...
21
1
import argparse import json from collections import OrderedDict import torch from huggingface_hub import cached_download, hf_hub_url from transformers import AutoImageProcessor, CvtConfig, CvtForImageClassification def lowerCAmelCase_ ( lowerCamelCase ): __magic_name__ : ...
21
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyInpaintPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.tex...
21
1
UpperCAmelCase_ : Optional[Any] = "\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n" Uppe...
21
import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device ...
21
1
def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase ): __magic_name__ : Tuple =word.split() def justify(lowerCamelCase , lowerCamelCase , lowerCamelCase ) -> str: __magic_name__ : Optional[Any] =max_width - width __...
21
import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ : Any = logging.get_logger(__name__) Upp...
21
1
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_gpta import GPTaTokenizer ...
21
import heapq def lowerCAmelCase_ ( lowerCamelCase ): __magic_name__ : list[list] =[] # for each node and his adjacency list add them and the rank of the node to queue # using heapq module the queue will be filled like a Priority Queue # heapq works with...
21
1
def lowerCAmelCase_ ( lowerCamelCase ): __magic_name__ : List[Any] =[0] * len(lowerCamelCase ) __magic_name__ : Optional[Any] =[] __magic_name__ : str =[1] * len(lowerCamelCase ) for values in graph.values(): for i ...
21
UpperCAmelCase_ : int = range(2, 20 + 1) UpperCAmelCase_ : Tuple = [10**k for k in range(ks[-1] + 1)] UpperCAmelCase_ : dict[int, dict[int, list[list[int]]]] = {} def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ...
21
1
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokeniz...
21
from typing import List from .keymap import KEYMAP, get_character def lowerCAmelCase_ ( lowerCamelCase ): def decorator(lowerCamelCase ): __magic_name__ : str =getattr(lowerCamelCase , """handle_key""" , [] ) handle += [key] s...
21
1
import unittest import numpy as np import torch from diffusers import KarrasVePipeline, KarrasVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class __A ( unittest.TestCase ...
21
import os import jsonlines import numpy as np from tqdm import tqdm UpperCAmelCase_ : Dict = 2048 UpperCAmelCase_ : int = 4096 UpperCAmelCase_ : Any = 42 UpperCAmelCase_ : Optional[int] = os.environ.pop("PROCESS_TRAIN", "false") UpperCA...
21
1
import torch from diffusers import KDPMaDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class __A ( UpperCamelCase__ ): UpperCamelCase = (KDPMaDiscreteScheduler,) UpperCamelCase = 10 ...
21
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ : Tuple = logging.get_logger(__name__) UpperCAmelCase_ : List[str] = { ...
21
1
import unittest import numpy as np import requests from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_...
21
from pathlib import Path import fire from tqdm import tqdm def lowerCAmelCase_ ( lowerCamelCase="ro" , lowerCamelCase="en" , lowerCamelCase="wmt16" , lowerCamelCase=None ): try: import datasets except (ModuleNotFoundError, ImportError): raise I...
21
1
import argparse import copy def lowerCAmelCase_ ( lowerCamelCase ): __magic_name__ : Tuple ={} with open(lowerCamelCase ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: __magic_name__ : Optiona...
21
from __future__ import annotations from fractions import Fraction def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase ): return ( num != den and num % 10 == den // 10 and (num // 10) / (den % 10) == num / den ) def lowerCAmelCase_ ( lowerCamelCase ...
21
1
import json import os import unittest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_token...
21
from collections.abc import Callable from math import pi, sqrt from random import uniform from statistics import mean def lowerCAmelCase_ ( lowerCamelCase ): # A local function to see if a dot lands in the circle. def is_in_circle(lowerCamelCase , lowerCamelCase ) -> bool:...
21
1
import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch ...
21
import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class __A ( tf.keras.layers.Layer ): def __init__( self ...
21
1
from dataclasses import asdict, dataclass from typing import Optional from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ : Union[str, Any] = logging.get_logger(__name__) # TODO Update this UpperCAmelCase_ : Optional[int] ...
21
import math import tensorflow as tf from packaging import version def lowerCAmelCase_ ( lowerCamelCase ): __magic_name__ : str =tf.convert_to_tensor(lowerCamelCase ) __magic_name__ : List[str] =0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0...
21
1
import math import tensorflow as tf from packaging import version def lowerCAmelCase_ ( lowerCamelCase ): __magic_name__ : str =tf.convert_to_tensor(lowerCamelCase ) __magic_name__ : List[str] =0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0...
21
from collections.abc import Sequence def lowerCAmelCase_ ( lowerCamelCase = None ): if nums is None or not nums: raise ValueError("""Input sequence should not be empty""" ) __magic_name__ : str =nums[0] for i in range(1 , len(lowerCamelCase ) ...
21
1
def lowerCAmelCase_ ( lowerCamelCase ): return " ".join(input_str.split()[::-1] ) if __name__ == "__main__": import doctest doctest.testmod()
21
from dataclasses import dataclass, field from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import pyarrow as pa if TYPE_CHECKING: from .features import FeatureType @dataclass class __A : UpperCamelCase = 42 UpperCamelCase ...
21
1
import json import os from typing import Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCAmelCase_ : Dict = logging.get_logger(__name__) UpperCAmelCase_ : List[Any] = { "vocab_file":...
21
from sklearn.metrics import matthews_corrcoef import datasets UpperCAmelCase_ : Dict = "\nCompute the Matthews correlation coefficient (MCC)\n\nThe Matthews correlation coefficient is used in machine learning as a\nmeasure of the quality of binary and multiclass classifications. It take...
21
1
def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase = False ): if n == 2: return True if not n % 2 or n < 2: return False if n > 5 and n % 10 not in (1, 3, 7, 9): # can quickly check last digit return False if n > 3317044064679887385961981 and not...
21
import importlib import torch import yaml from omegaconf import OmegaConf from taming.models.vqgan import VQModel def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase=False ): __magic_name__ : Optional[int] =OmegaConf.load(lowerCamelCase ) if display: ...
21
1
from __future__ import annotations def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase ): if b == 0: return (1, 0) ((__magic_name__) , (__magic_name__)) : Optional[int] =extended_euclid(lowerCamelCase , a % b ) __magic_name__ ...
21
import unittest from accelerate import debug_launcher from accelerate.test_utils import require_cpu, test_ops, test_script @require_cpu class __A ( unittest.TestCase ): def A__ ( self :Tuple ): '''simple docstring''' debug_launcher(test_s...
21
1
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_bart import BartToken...
21
UpperCAmelCase_ : Tuple = 0 # The first color of the flag. UpperCAmelCase_ : Any = 1 # The second color of the flag. UpperCAmelCase_ : str = 2 # The third color of the flag. UpperCAmelCase_ : Tuple = (red, white, blue) def lowerCAmel...
21
1
from __future__ import annotations import unittest from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import Conf...
21
# DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from typing import Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import randn_tensor from .scheduling_utils import SchedulerMixin ...
21
1
import os from typing import Dict, List, Tuple, TypeVar, Union UpperCAmelCase_ : str = TypeVar("T") UpperCAmelCase_ : Optional[int] = Union[List[T], Tuple[T, ...]] UpperCAmelCase_ : Dict = Union[T, List[T], Dict[str, T]] UpperCAmelCase_ : Any ...
21
from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy UpperCAmelCase_ : Dict = logging.get_logger(_...
21
1
import re import string from collections import Counter import sacrebleu import sacremoses from packaging import version import datasets UpperCAmelCase_ : List[Any] = "\n@inproceedings{xu-etal-2016-optimizing,\n title = {Optimizing Statistical Machine Translation for Text Simp...
21
import inspect import unittest import torch import torch.nn as nn from accelerate.hooks import ( AlignDevicesHook, ModelHook, SequentialHook, add_hook_to_module, attach_align_device_hook, remove_hook_from_module, remove_hook_from_submodules, ) from accelerate.test_...
21
1
from typing import List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ : Tuple = logging.get_logger(__name__) UpperCAmelCase_ : List[str] = { "huggingface/autoformer-tourism-monthly": "https://huggingfac...
21
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyInpaintPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.tex...
21
1
import json import logging import math import os import sys from dataclasses import dataclass, field from typing import Optional from datasets import Dataset, load_dataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_FOR_MASKED_LM_MAPPING, AutoConfig, A...
21
import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device ...
21
1
import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature fro...
21
import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ : Any = logging.get_logger(__name__) Upp...
21
1
import numpy as np from matplotlib import pyplot as plt from sklearn.datasets import load_iris from sklearn.metrics import ConfusionMatrixDisplay from sklearn.model_selection import train_test_split from xgboost import XGBClassifier def lowerCAmelCase_ ( lowerCamelCase ): return (dat...
21
import heapq def lowerCAmelCase_ ( lowerCamelCase ): __magic_name__ : list[list] =[] # for each node and his adjacency list add them and the rank of the node to queue # using heapq module the queue will be filled like a Priority Queue # heapq works with...
21
1
from scipy.stats import spearmanr import datasets UpperCAmelCase_ : Optional[Any] = "\nThe Spearman rank-order correlation coefficient is a measure of the\nrelationship between two datasets. Like other correlation coefficients,\nthis one varies between -1 and +1 with 0 implying no corre...
21
UpperCAmelCase_ : int = range(2, 20 + 1) UpperCAmelCase_ : Tuple = [10**k for k in range(ks[-1] + 1)] UpperCAmelCase_ : dict[int, dict[int, list[list[int]]]] = {} def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ...
21
1
# A Bipartite Graph is a graph whose vertices can be divided into two independent sets, # U and V such that every edge (u, v) either connects a vertex from U to V or a vertex # from V to U. In other words, for every edge (u, v), either u belongs to U and v to V, # or u belongs to V and v to U. We can also ...
21
from typing import List from .keymap import KEYMAP, get_character def lowerCAmelCase_ ( lowerCamelCase ): def decorator(lowerCamelCase ): __magic_name__ : str =getattr(lowerCamelCase , """handle_key""" , [] ) handle += [key] s...
21
1
from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_herbert import HerbertTokenizer UpperCAmelCase_ : List[Any] = logging.get_logger(__name__) UpperCAmelCase_ : List[str] ...
21
import os import jsonlines import numpy as np from tqdm import tqdm UpperCAmelCase_ : Dict = 2048 UpperCAmelCase_ : int = 4096 UpperCAmelCase_ : Any = 42 UpperCAmelCase_ : Optional[int] = os.environ.pop("PROCESS_TRAIN", "false") UpperCA...
21
1
import json import os import tempfile import transformers import datasets from utils import generate_example_dataset, get_duration UpperCAmelCase_ : List[str] = 500000 UpperCAmelCase_ , UpperCAmelCase_ : List[Any] = os.path.split(__file__) UpperCAmelCase_ :...
21
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ : Tuple = logging.get_logger(__name__) UpperCAmelCase_ : List[str] = { ...
21
1
import unittest import numpy as np from transformers import RobertaPreLayerNormConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available...
21
from pathlib import Path import fire from tqdm import tqdm def lowerCAmelCase_ ( lowerCamelCase="ro" , lowerCamelCase="en" , lowerCamelCase="wmt16" , lowerCamelCase=None ): try: import datasets except (ModuleNotFoundError, ImportError): raise I...
21
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_ind...
21
from __future__ import annotations from fractions import Fraction def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase ): return ( num != den and num % 10 == den // 10 and (num // 10) / (den % 10) == num / den ) def lowerCAmelCase_ ( lowerCamelCase ...
21
1
def lowerCAmelCase_ ( lowerCamelCase = 10 ): if not isinstance(lowerCamelCase , lowerCamelCase ) or n < 0: raise ValueError("""Invalid input""" ) __magic_name__ : List[Any] =10**n __magic_name__ : List[str] =28433 * (pow(2 , 7830...
21
from collections.abc import Callable from math import pi, sqrt from random import uniform from statistics import mean def lowerCAmelCase_ ( lowerCamelCase ): # A local function to see if a dot lands in the circle. def is_in_circle(lowerCamelCase , lowerCamelCase ) -> bool:...
21
1
import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class __A ( UpperCamelCase__ ): UpperCamelCase = (PNDMScheduler,) UpperCamelCase = (("""num_inference_steps""", 50),) d...
21
import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class __A ( tf.keras.layers.Layer ): def __init__( self ...
21
1
import json import os from datetime import date from pathlib import Path from tabulate import DataRow, TableFormat, tabulate UpperCAmelCase_ : List[Any] = TableFormat( lineabove=None, linebelowheader=None, linebetweenrows=None, linebelow=None, headerrow=DataR...
21
import math import tensorflow as tf from packaging import version def lowerCAmelCase_ ( lowerCamelCase ): __magic_name__ : str =tf.convert_to_tensor(lowerCamelCase ) __magic_name__ : List[str] =0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0...
21
1
import requests UpperCAmelCase_ : List[str] = "https://newsapi.org/v1/articles?source=bbc-news&sortBy=top&apiKey=" def lowerCAmelCase_ ( lowerCamelCase ): # fetching a list of articles in json format __magic_name__ : List[Any] =requests.get(_NEWS_...
21
from collections.abc import Sequence def lowerCAmelCase_ ( lowerCamelCase = None ): if nums is None or not nums: raise ValueError("""Input sequence should not be empty""" ) __magic_name__ : str =nums[0] for i in range(1 , len(lowerCamelCase ) ...
21
1
from sklearn.metrics import matthews_corrcoef import datasets UpperCAmelCase_ : Dict = "\nCompute the Matthews correlation coefficient (MCC)\n\nThe Matthews correlation coefficient is used in machine learning as a\nmeasure of the quality of binary and multiclass classifications. It take...
21
from dataclasses import dataclass, field from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import pyarrow as pa if TYPE_CHECKING: from .features import FeatureType @dataclass class __A : UpperCamelCase = 42 UpperCamelCase ...
21
1
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Value from .base import TaskTemplate @dataclass(frozen=UpperCamelCase__ ) class __A ( UpperCamelCase__ ): # `task` is not a ClassVar sin...
21
from sklearn.metrics import matthews_corrcoef import datasets UpperCAmelCase_ : Dict = "\nCompute the Matthews correlation coefficient (MCC)\n\nThe Matthews correlation coefficient is used in machine learning as a\nmeasure of the quality of binary and multiclass classifications. It take...
21
1
import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers...
21
import importlib import torch import yaml from omegaconf import OmegaConf from taming.models.vqgan import VQModel def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase=False ): __magic_name__ : Optional[int] =OmegaConf.load(lowerCamelCase ) if display: ...
21
1
import argparse import torch from torch import nn from transformers import SpeechaTextConfig, SpeechaTextForConditionalGeneration def lowerCAmelCase_ ( lowerCamelCase ): __magic_name__ : int =[ """encoder.version""", """decoder.version""", ...
21
import unittest from accelerate import debug_launcher from accelerate.test_utils import require_cpu, test_ops, test_script @require_cpu class __A ( unittest.TestCase ): def A__ ( self :Tuple ): '''simple docstring''' debug_launcher(test_s...
21
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCAmelCase_ : int = { "configuration_bloom": ["BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP", "BloomConfig", "BloomOnnxConfig"], } try: ...
21
UpperCAmelCase_ : Tuple = 0 # The first color of the flag. UpperCAmelCase_ : Any = 1 # The second color of the flag. UpperCAmelCase_ : str = 2 # The third color of the flag. UpperCAmelCase_ : Tuple = (red, white, blue) def lowerCAmel...
21
1
def lowerCAmelCase_ ( lowerCamelCase ): return "".join([hex(lowerCamelCase )[2:].zfill(2 ).upper() for byte in list(lowerCamelCase )] ) def lowerCAmelCase_ ( lowerCamelCase ): # Check data validity, following RFC3548 # https://www.ietf.org/rfc/rfc3548.txt if (len(low...
21
# DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from typing import Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import randn_tensor from .scheduling_utils import SchedulerMixin ...
21
1
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from transformers import TensorFlowBenchmark, TensorFlowBenchmar...
21
from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy UpperCAmelCase_ : Dict = logging.get_logger(_...
21
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase_ : Any = { "configuration_blenderbot_small": [ ...
21
import inspect import unittest import torch import torch.nn as nn from accelerate.hooks import ( AlignDevicesHook, ModelHook, SequentialHook, add_hook_to_module, attach_align_device_hook, remove_hook_from_module, remove_hook_from_submodules, ) from accelerate.test_...
21
1
import unittest import numpy as np from diffusers import LMSDiscreteScheduler, OnnxStableDiffusionInpaintPipeline from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common ...
21
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyInpaintPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.tex...
21
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase_ : Dict = {"configuration_fnet": ["FNET_PRETRAINED_CONFIG_ARCHIVE_MAP"...
21
import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device ...
21
1
UpperCAmelCase_ : int = "Input must be a string of 8 numbers plus letter" UpperCAmelCase_ : Tuple = "TRWAGMYFPDXBNJZSQVHLCKE" def lowerCAmelCase_ ( lowerCamelCase ): if not isinstance(lowerCamelCase , lowerCamelCase ): __magic_name__ : ...
21
import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ : Any = logging.get_logger(__name__) Upp...
21
1
import os import time import warnings from dataclasses import dataclass, field from enum import Enum from typing import List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...tokenization_utils_base import PreTrainedTokenizerBase from ...uti...
21
import heapq def lowerCAmelCase_ ( lowerCamelCase ): __magic_name__ : list[list] =[] # for each node and his adjacency list add them and the rank of the node to queue # using heapq module the queue will be filled like a Priority Queue # heapq works with...
21
1
import unittest from transformers import RoFormerTokenizer, RoFormerTokenizerFast from transformers.testing_utils import require_rjieba, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_rjieba @require_tokenizers class __A ( UpperCamelCase__ ...
21
UpperCAmelCase_ : int = range(2, 20 + 1) UpperCAmelCase_ : Tuple = [10**k for k in range(ks[-1] + 1)] UpperCAmelCase_ : dict[int, dict[int, list[list[int]]]] = {} def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ...
21
1
import argparse from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection from diffusers import UnCLIPImageVariationPipeline, UnCLIPPipeline if __name__ == "__main__": UpperCAmelCase_ : Optional[Any] = argparse.ArgumentParser() parser.add_argument("...
21
from typing import List from .keymap import KEYMAP, get_character def lowerCAmelCase_ ( lowerCamelCase ): def decorator(lowerCamelCase ): __magic_name__ : str =getattr(lowerCamelCase , """handle_key""" , [] ) handle += [key] s...
21
1
from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import TensorType, logging if TYPE_CHECKING: from ...onnx.config import PatchingSpec ...
21
import os import jsonlines import numpy as np from tqdm import tqdm UpperCAmelCase_ : Dict = 2048 UpperCAmelCase_ : int = 4096 UpperCAmelCase_ : Any = 42 UpperCAmelCase_ : Optional[int] = os.environ.pop("PROCESS_TRAIN", "false") UpperCA...
21
1
import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.data imp...
21
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ : Tuple = logging.get_logger(__name__) UpperCAmelCase_ : List[str] = { ...
21
1
import argparse import json import os import fairseq import torch from torch import nn from transformers import ( SpeechaTextaConfig, SpeechaTextaForCausalLM, SpeechaTextaTokenizer, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeat...
21
from pathlib import Path import fire from tqdm import tqdm def lowerCAmelCase_ ( lowerCamelCase="ro" , lowerCamelCase="en" , lowerCamelCase="wmt16" , lowerCamelCase=None ): try: import datasets except (ModuleNotFoundError, ImportError): raise I...
21
1
from __future__ import annotations def lowerCAmelCase_ ( lowerCamelCase ): if len(lowerCamelCase ) < 2: raise ValueError("""Monogons and Digons are not polygons in the Euclidean space""" ) if any(i <= 0 for i in nums ): raise ValueError("""All values must be greater...
21
from __future__ import annotations from fractions import Fraction def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase ): return ( num != den and num % 10 == den // 10 and (num // 10) / (den % 10) == num / den ) def lowerCAmelCase_ ( lowerCamelCase ...
21
1
def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase ): print("""\nThe shortest path matrix using Floyd Warshall algorithm\n""" ) for i in range(lowerCamelCase ): for j in range(lowerCamelCase ): if dist[i][j] != float("""inf""" ): print(int(dist[...
21
from collections.abc import Callable from math import pi, sqrt from random import uniform from statistics import mean def lowerCAmelCase_ ( lowerCamelCase ): # A local function to see if a dot lands in the circle. def is_in_circle(lowerCamelCase , lowerCamelCase ) -> bool:...
21
1
import json import sys def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase ): with open(lowerCamelCase , encoding="""utf-8""" ) as f: __magic_name__ : Tuple =json.load(lowerCamelCase ) __magic_name__ : Optional[Any] =[""...
21
import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class __A ( tf.keras.layers.Layer ): def __init__( self ...
21
1
from typing import Callable, List, Optional, Tuple, Union import torch from transformers import CLIPTextModel, CLIPTokenizer from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin, TransformeraDModel, VQModel from ...schedulers import VQDiffusionScheduler ...
21
import math import tensorflow as tf from packaging import version def lowerCAmelCase_ ( lowerCamelCase ): __magic_name__ : str =tf.convert_to_tensor(lowerCamelCase ) __magic_name__ : List[str] =0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0...
21
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCAmelCase_ : Union[str, Any] = { "configuration_conditional_detr": [ "CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP", "...
21
from collections.abc import Sequence def lowerCAmelCase_ ( lowerCamelCase = None ): if nums is None or not nums: raise ValueError("""Input sequence should not be empty""" ) __magic_name__ : str =nums[0] for i in range(1 , len(lowerCamelCase ) ...
21
1
import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGeneration, ...
21
from dataclasses import dataclass, field from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import pyarrow as pa if TYPE_CHECKING: from .features import FeatureType @dataclass class __A : UpperCamelCase = 42 UpperCamelCase ...
21
1
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from trans...
21
from sklearn.metrics import matthews_corrcoef import datasets UpperCAmelCase_ : Dict = "\nCompute the Matthews correlation coefficient (MCC)\n\nThe Matthews correlation coefficient is used in machine learning as a\nmeasure of the quality of binary and multiclass classifications. It take...
21
1
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ : Tuple = logging.get_logger(__name__) UpperCAmelCase_ : Tuple = { "uw-madison/mra-base-512-4": "https://huggingface.co/uw-madison/mra-base-512-4/resolve/main/config.jso...
21
import importlib import torch import yaml from omegaconf import OmegaConf from taming.models.vqgan import VQModel def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase=False ): __magic_name__ : Optional[int] =OmegaConf.load(lowerCamelCase ) if display: ...
21
1
import os import zipfile import pytest from datasets.utils.extract import ( BzipaExtractor, Extractor, GzipExtractor, LzaExtractor, SevenZipExtractor, TarExtractor, XzExtractor, ZipExtractor, ZstdExtractor, ) from .utils import require_lza, require_pyazr...
21
import unittest from accelerate import debug_launcher from accelerate.test_utils import require_cpu, test_ops, test_script @require_cpu class __A ( unittest.TestCase ): def A__ ( self :Tuple ): '''simple docstring''' debug_launcher(test_s...
21
1
from __future__ import annotations def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase = None , lowerCamelCase = None ): if start is None: __magic_name__ : Tuple =0 if end is None: __magic_name__ : Tuple =len(low...
21
UpperCAmelCase_ : Tuple = 0 # The first color of the flag. UpperCAmelCase_ : Any = 1 # The second color of the flag. UpperCAmelCase_ : str = 2 # The third color of the flag. UpperCAmelCase_ : Tuple = (red, white, blue) def lowerCAmel...
21
1
from dataclasses import dataclass from typing import Dict, Optional, Union import torch import torch.nn.functional as F from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .attention...
21
# DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from typing import Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import randn_tensor from .scheduling_utils import SchedulerMixin ...
21
1
import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @requ...
21
from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy UpperCAmelCase_ : Dict = logging.get_logger(_...
21
1
import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterM...
21
import inspect import unittest import torch import torch.nn as nn from accelerate.hooks import ( AlignDevicesHook, ModelHook, SequentialHook, add_hook_to_module, attach_align_device_hook, remove_hook_from_module, remove_hook_from_submodules, ) from accelerate.test_...
21
1
import colorsys from PIL import Image # type: ignore def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ): __magic_name__ : Any =x __magic_name__ : Dict =y for step in range(lowerCamelCase ): # noqa: B007...
21
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyInpaintPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.tex...
21
1
from typing import List from .keymap import KEYMAP, get_character def lowerCAmelCase_ ( lowerCamelCase ): def decorator(lowerCamelCase ): __magic_name__ : str =getattr(lowerCamelCase , """handle_key""" , [] ) handle += [key] s...
21
import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device ...
21
1
import inspect import unittest import torch import torch.nn as nn from accelerate.hooks import ( AlignDevicesHook, ModelHook, SequentialHook, add_hook_to_module, attach_align_device_hook, remove_hook_from_module, remove_hook_from_submodules, ) from accelerate.test_...
21
import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ : Any = logging.get_logger(__name__) Upp...
21
1
import pyarrow.parquet as pq import pytest from datasets import Audio, Dataset, DatasetDict, Features, NamedSplit, Sequence, Value, config from datasets.features.image import Image from datasets.io.parquet import ParquetDatasetReader, ParquetDatasetWriter, get_writer_batch_size from ..utils import as...
21
import heapq def lowerCAmelCase_ ( lowerCamelCase ): __magic_name__ : list[list] =[] # for each node and his adjacency list add them and the rank of the node to queue # using heapq module the queue will be filled like a Priority Queue # heapq works with...
21
1
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_download, hf_hub_url from PIL import Image from transformers import DetaConfig, DetaForObjectDetection, DetaImageProcessor, SwinConfig from transformers.utils impor...
21
UpperCAmelCase_ : int = range(2, 20 + 1) UpperCAmelCase_ : Tuple = [10**k for k in range(ks[-1] + 1)] UpperCAmelCase_ : dict[int, dict[int, list[list[int]]]] = {} def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase , lowerCamelCase ...
21
1
from numpy import exp, pi, sqrt def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase = 0.0 , lowerCamelCase = 1.0 ): return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod(...
21
from typing import List from .keymap import KEYMAP, get_character def lowerCAmelCase_ ( lowerCamelCase ): def decorator(lowerCamelCase ): __magic_name__ : str =getattr(lowerCamelCase , """handle_key""" , [] ) handle += [key] s...
21
1
from argparse import ArgumentParser, Namespace from ..utils import logging from . import BaseTransformersCLICommand def lowerCAmelCase_ ( lowerCamelCase ): return ConvertCommand( args.model_type , args.tf_checkpoint , args.pytorch_dump_output , args.config ...
21
import os import jsonlines import numpy as np from tqdm import tqdm UpperCAmelCase_ : Dict = 2048 UpperCAmelCase_ : int = 4096 UpperCAmelCase_ : Any = 42 UpperCAmelCase_ : Optional[int] = os.environ.pop("PROCESS_TRAIN", "false") UpperCA...
21
1
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTImageProcessor, ViTMSNConfig, ViTMSNModel from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD torch.set_grad_enabl...
21
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ : Tuple = logging.get_logger(__name__) UpperCAmelCase_ : List[str] = { ...
21
1
import heapq as hq import math from collections.abc import Iterator class __A : def __init__( self :Optional[Any] , __snake_case :Optional[Any] ): '''simple docstring''' __magic_name__ : int =str(id_ ) __magic_name__ ...
21
from pathlib import Path import fire from tqdm import tqdm def lowerCAmelCase_ ( lowerCamelCase="ro" , lowerCamelCase="en" , lowerCamelCase="wmt16" , lowerCamelCase=None ): try: import datasets except (ModuleNotFoundError, ImportError): raise I...
21
1
def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase ): __magic_name__ : Optional[int] =1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): __magic_name__ : Union[str, Any] =n - k # Calculate C(n,k) ...
21
from __future__ import annotations from fractions import Fraction def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase ): return ( num != den and num % 10 == den // 10 and (num // 10) / (den % 10) == num / den ) def lowerCAmelCase_ ( lowerCamelCase ...
21
1
from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record UpperCAmelCase_ : int = "\\n@article{wang2019superglue,\n title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems},\n ...
21
from collections.abc import Callable from math import pi, sqrt from random import uniform from statistics import mean def lowerCAmelCase_ ( lowerCamelCase ): # A local function to see if a dot lands in the circle. def is_in_circle(lowerCamelCase , lowerCamelCase ) -> bool:...
21
1
from __future__ import annotations def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase ): __magic_name__ : Tuple =sorted(numsa + numsa ) __magic_name__ , __magic_name__ : Optional[Any] =divmod(len(lowerCamelCase ) , 2 ) i...
21
import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class __A ( tf.keras.layers.Layer ): def __init__( self ...
21
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ : Any = logging.get_logger(__name__) UpperCAmelCase_ : ...
21
import math import tensorflow as tf from packaging import version def lowerCAmelCase_ ( lowerCamelCase ): __magic_name__ : str =tf.convert_to_tensor(lowerCamelCase ) __magic_name__ : List[str] =0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0...
21
1
from copy import deepcopy import torch import torch.nn.functional as F from torch.optim import AdamW from torch.optim.lr_scheduler import LambdaLR from torch.utils.data import DataLoader from accelerate.accelerator import Accelerator from accelerate.state import GradientState from accelerate.test_...
21
from collections.abc import Sequence def lowerCAmelCase_ ( lowerCamelCase = None ): if nums is None or not nums: raise ValueError("""Input sequence should not be empty""" ) __magic_name__ : str =nums[0] for i in range(1 , len(lowerCamelCase ) ...
21
1
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_...
21
from dataclasses import dataclass, field from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import pyarrow as pa if TYPE_CHECKING: from .features import FeatureType @dataclass class __A : UpperCamelCase = 42 UpperCamelCase ...
21
1
# this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won't be considered...
21
from sklearn.metrics import matthews_corrcoef import datasets UpperCAmelCase_ : Dict = "\nCompute the Matthews correlation coefficient (MCC)\n\nThe Matthews correlation coefficient is used in machine learning as a\nmeasure of the quality of binary and multiclass classifications. It take...
21
1
from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy UpperCAmelCase_ : Dict = logging.get_logger(_...
21
import importlib import torch import yaml from omegaconf import OmegaConf from taming.models.vqgan import VQModel def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase=False ): __magic_name__ : Optional[int] =OmegaConf.load(lowerCamelCase ) if display: ...
21
1