code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
import inspect import warnings from typing import Any, Dict, Optional, Union from packaging import version def _snake_case ( *__snake_case , __snake_case = None , __snake_case=True , __snake_case=2 ): from .. import __version__ _UpperCamelCase = take_from _UpperCame...
10
from __future__ import annotations import math import numpy as np from numpy.linalg import norm def _snake_case ( __snake_case , __snake_case ): return math.sqrt(sum(pow(a - b , 2 ) for a, b in zip(__snake_case , __snake_case ) ) ) def _snake_case ( __snake_cas...
10
1
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto.configuration_auto import CONFIG_MAPPING _lowerCAmelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( __lowercase ): UpperCAmelCase = "upernet" def __...
10
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow f...
10
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCAmelCase = { "configuration_git": ["GIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "GitConfig", "GitVisionConfig"], "processing_git": ["GitProcessor"], } try: if not is_torch_availab...
10
import random import torch from huggingface_hub import HfApi from diffusers import UNetaDModel _lowerCAmelCase = HfApi() _lowerCAmelCase = {} # fmt: off _lowerCAmelCase = torch.tensor([ -0.7515, -1.6883, 0.2420, 0.0300, 0.6347, 1.3433, -1.1743, -3.7467, 1.2342, -2.2485, 0.4636, 0.8076,...
10
1
from __future__ import annotations _lowerCAmelCase = 1.6021E-19 # units = C def _snake_case ( __snake_case , __snake_case , __snake_case , ): if (conductivity, electron_conc, mobility).count(0 ) != 1: raise ValueError('''You cannot supply more or less than 2 values''' ...
10
from typing import List from .keymap import KEYMAP, get_character def _snake_case ( __snake_case ): def decorator(__snake_case ): _UpperCamelCase = getattr(__snake_case , '''handle_key''' , [] ) handle += [key] setattr(__snake_case , ''...
10
1
import requests _lowerCAmelCase = "YOUR API KEY" def _snake_case ( __snake_case , __snake_case = giphy_api_key ): _UpperCamelCase = '''+'''.join(query.split() ) _UpperCamelCase = f"""https://api.giphy.com/v1/gifs/search?q={formatted_query}&api_key={api_key}""...
10
import unittest from transformers import ( MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TextGenerationPipeline, logging, pipeline, ) from transformers.testing_utils import ( CaptureLogger, is_pipeline_test, require_accelerate, require_tf, require_torch, requi...
10
1
from __future__ import annotations from typing import Any def _snake_case ( __snake_case ): if not postfix_notation: return 0 _UpperCamelCase = {'''+''', '''-''', '''*''', '''/'''} _UpperCamelCase = [] for token in postfix_notation: if token in...
10
def _snake_case ( __snake_case = 100 ): _UpperCamelCase = (n * (n + 1) // 2) ** 2 _UpperCamelCase = n * (n + 1) * (2 * n + 1) // 6 return sum_cubes - sum_squares if __name__ == "__main__": print(f'{solution() = }')
10
1
# coding=utf-8 # Copyright 2020 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
10
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDAR...
10
1
import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa imp...
10
import os import re import shutil import sys import tempfile import unittest import black _lowerCAmelCase = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, "utils")) import check_copies # noqa: E402 # This is the reference code that ...
10
1
from __future__ import annotations import unittest from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...t...
10
import json import logging import os import re import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import datasets import numpy as np import torch import torchaudio from packaging import version from torch import nn import transformers from transformers import ( ...
10
1
def _snake_case ( __snake_case , __snake_case , __snake_case ): return not any( neighbour == 1 and colored_vertices[i] == color for i, neighbour in enumerate(__snake_case ) ) def _snake_case ( __snake_case , __snake_case , __snake_case , __snake...
10
import math class lowerCAmelCase_ : def __init__( self : Tuple , _A : int=0 ): # a graph with Node 0,1,...,N-1 _UpperCamelCase = n _UpperCamelCase = [ [math.inf for j in range(0 , _A )] for i in range(0 , _A...
10
1
_lowerCAmelCase = "0.21.0" from .accelerator import Accelerator from .big_modeling import ( cpu_offload, cpu_offload_with_hook, disk_offload, dispatch_model, init_empty_weights, init_on_device, load_checkpoint_and_dispatch, ) from .data_loader import skip_first_batches from .launcher...
10
import dataclasses import json import warnings from dataclasses import dataclass, field from time import time from typing import List from ..utils import logging _lowerCAmelCase = logging.get_logger(__name__) def _snake_case ( __snake_case=None , __snake_case=None ): return field(default_...
10
1
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def _snake_case ( ): _UpperCamelCase = [randint(-1000 , 1000 ) for i in range(10 )] _UpperCamelCase = randint(-5000 , 5000 ) ret...
10
import inspect import warnings from typing import Any, Dict, Optional, Union from packaging import version def _snake_case ( *__snake_case , __snake_case = None , __snake_case=True , __snake_case=2 ): from .. import __version__ _UpperCamelCase = take_from _UpperCame...
10
1
from typing import List, Optional, Union import numpy as np import tensorflow as tf from .utils import logging _lowerCAmelCase = logging.get_logger(__name__) def _snake_case ( __snake_case ): if isinstance(__snake_case , np.ndarray ): return list(tensor.shape ) ...
10
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCo...
10
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCAmelCase = { "configuration_pegasus_x": ["PEGASUS_X_PRETRAINED_CONFIG_ARCHIVE_MAP", "PegasusXConfig"], } try: if not is_torch_available(): raise OptionalDependencyNo...
10
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase = logging.get_logger(__name__) _lowerCAmelCase = { "microsoft/trocr-base-handwritten": ( "https://huggingface.co/microsoft/trocr-base-handwritten/resolve/main/config.json" ), # See all T...
10
1
import numpy as np import torch from torch.utils.data import Dataset, IterableDataset from ..utils.generic import ModelOutput class lowerCAmelCase_ ( __lowercase ): def __init__( self : int , _A : Optional[int] , _A : Dict , _A : Union[str, Any] ...
10
import os import tempfile import unittest from transformers import FlaubertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, rand...
10
1
import json import os import unittest from transformers import BatchEncoding, LEDTokenizer, LEDTokenizerFast from transformers.models.led.tokenization_led import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_token...
10
from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTes...
10
1
import random def _snake_case ( __snake_case , __snake_case , __snake_case = False ): _UpperCamelCase = {i: [] for i in range(__snake_case )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_gr...
10
from __future__ import annotations _lowerCAmelCase = [True] * 1_000_001 _lowerCAmelCase = 2 while i * i <= 1_000_000: if seive[i]: for j in range(i * i, 1_000_001, i): _lowerCAmelCase = False i += 1 def _snake_case ( __snake_case ): return seive[n...
10
1
import inspect import tempfile from collections import OrderedDict, UserDict from collections.abc import MutableMapping from contextlib import ExitStack, contextmanager from dataclasses import fields from enum import Enum from typing import Any, ContextManager, List, Tuple import numpy as np from .import_utils import...
10
import unittest from transformers import DebertaVaTokenizer, DebertaVaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin _lowerCAmelCase = get_tests_dir("fixtures/spiece.model") @requ...
10
1
from __future__ import annotations _lowerCAmelCase = [True] * 1_000_001 _lowerCAmelCase = 2 while i * i <= 1_000_000: if seive[i]: for j in range(i * i, 1_000_001, i): _lowerCAmelCase = False i += 1 def _snake_case ( __snake_case ): return seive[n...
10
import sys from collections import defaultdict class lowerCAmelCase_ : def __init__( self : Optional[int] ): _UpperCamelCase = [] def UpperCamelCase_ ( self : Any , _A : str ): return self.node_position[verte...
10
1
import webbrowser from sys import argv from urllib.parse import parse_qs, quote import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": _lowerCAmelCase = "%20".join(argv[1:]) if len(argv) > 1 else quote(str(input("Search: "))) print("Googling.......
10
import logging import os from .state import PartialState class lowerCAmelCase_ ( logging.LoggerAdapter ): @staticmethod def UpperCamelCase_ ( _A : Any ): _UpperCamelCase = PartialState() return not main_process_only or (main_proces...
10
1
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDAR...
10
import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _lowerCAmelCase = "▁" _lowerCAmelCase ...
10
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging _lowerCAmelCase = logging.get_logger(__name__) _lowerCAmelCase = "▁" _lo...
10
import gc import unittest import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DDPMScheduler, PriorTransformer, StableUnCLIPPipeline, UNetaDConditionModel, ) from diffusers.pipeline...
10
1
import os import sys from contextlib import contextmanager # Windows only if os.name == "nt": import ctypes import msvcrt # noqa class lowerCAmelCase_ ( ctypes.Structure ): # _fields is a specific attr expected by ctypes UpperCAmelCase = [("size", ctypes...
10
from __future__ import annotations import math import numpy as np from numpy.linalg import norm def _snake_case ( __snake_case , __snake_case ): return math.sqrt(sum(pow(a - b , 2 ) for a, b in zip(__snake_case , __snake_case ) ) ) def _snake_case ( __snake_cas...
10
1
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING _lowerCAmelCase = logging.get_logger(__name__) _lowerCAmelCase = ...
10
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow f...
10
1
# NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from ...utils import deprecate from ..controlnet.pipeline_flax_controlnet import FlaxStableDiffusionControlNetPipeline # noqa: F401 deprecate( "...
10
import random import torch from huggingface_hub import HfApi from diffusers import UNetaDModel _lowerCAmelCase = HfApi() _lowerCAmelCase = {} # fmt: off _lowerCAmelCase = torch.tensor([ -0.7515, -1.6883, 0.2420, 0.0300, 0.6347, 1.3433, -1.1743, -3.7467, 1.2342, -2.2485, 0.4636, 0.8076,...
10
1
def _snake_case ( __snake_case ): _UpperCamelCase = [0] * len(__snake_case ) for i in range(1 , len(__snake_case ) ): # use last results for better performance - dynamic programming _UpperCamelCase = prefix_result[i - 1] while...
10
from typing import List from .keymap import KEYMAP, get_character def _snake_case ( __snake_case ): def decorator(__snake_case ): _UpperCamelCase = getattr(__snake_case , '''handle_key''' , [] ) handle += [key] setattr(__snake_case , ''...
10
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _lowerCAmelCase = { "configuration_rag": ["RagConfig"], "retrieval_rag": ["RagRetriever"], "tokenization_rag": ["RagTokenizer"], } try: if not is_torch_avail...
10
import unittest from transformers import ( MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TextGenerationPipeline, logging, pipeline, ) from transformers.testing_utils import ( CaptureLogger, is_pipeline_test, require_accelerate, require_tf, require_torch, requi...
10
1
def _snake_case ( __snake_case , __snake_case ): print('''\nThe shortest path matrix using Floyd Warshall algorithm\n''' ) for i in range(__snake_case ): for j in range(__snake_case ): if dist[i][j] != float('''inf''' ): print(in...
10
def _snake_case ( __snake_case = 100 ): _UpperCamelCase = (n * (n + 1) // 2) ** 2 _UpperCamelCase = n * (n + 1) * (2 * n + 1) // 6 return sum_cubes - sum_squares if __name__ == "__main__": print(f'{solution() = }')
10
1
from typing import List import datasets from datasets.tasks import AudioClassification from ..folder_based_builder import folder_based_builder _lowerCAmelCase = datasets.utils.logging.get_logger(__name__) class lowerCAmelCase_ ( folder_based_builder.FolderBasedBuilderConfig ): UpperCAmel...
10
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDAR...
10
1
from __future__ import annotations import unittest from transformers import DebertaVaConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...
10
import os import re import shutil import sys import tempfile import unittest import black _lowerCAmelCase = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, "utils")) import check_copies # noqa: E402 # This is the reference code that ...
10
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase = logging.get_logger(__name__) _lowerCAmelCase = { "facebook/s2t-wav2vec2-large-en-de": ( "https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/config.json" ), # See all...
10
import json import logging import os import re import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import datasets import numpy as np import torch import torchaudio from packaging import version from torch import nn import transformers from transformers import ( ...
10
1
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class ...
10
import math class lowerCAmelCase_ : def __init__( self : Tuple , _A : int=0 ): # a graph with Node 0,1,...,N-1 _UpperCamelCase = n _UpperCamelCase = [ [math.inf for j in range(0 , _A )] for i in range(0 , _A...
10
1
from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...
10
import dataclasses import json import warnings from dataclasses import dataclass, field from time import time from typing import List from ..utils import logging _lowerCAmelCase = logging.get_logger(__name__) def _snake_case ( __snake_case=None , __snake_case=None ): return field(default_...
10
1
import argparse import json import os from collections import OrderedDict import torch from transformers import LukeConfig, LukeForMaskedLM, MLukeTokenizer, XLMRobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def _snake_case ( __snake_case , __snake_case , ...
10
import inspect import warnings from typing import Any, Dict, Optional, Union from packaging import version def _snake_case ( *__snake_case , __snake_case = None , __snake_case=True , __snake_case=2 ): from .. import __version__ _UpperCamelCase = take_from _UpperCame...
10
1
def _snake_case ( __snake_case , __snake_case ): return "\n".join( f"""{number} * {i} = {number * i}""" for i in range(1 , number_of_terms + 1 ) ) if __name__ == "__main__": print(multiplication_table(number=5, number_of_terms=10))
10
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCo...
10
1
class lowerCAmelCase_ : def __init__( self : Optional[Any] , _A : list ): _UpperCamelCase = set_counts _UpperCamelCase = max(_A ) _UpperCamelCase = len(_A ) _UpperCamelCase = [1] * ...
10
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase = logging.get_logger(__name__) _lowerCAmelCase = { "microsoft/trocr-base-handwritten": ( "https://huggingface.co/microsoft/trocr-base-handwritten/resolve/main/config.json" ), # See all T...
10
1
import os def _snake_case ( __snake_case = "input.txt" ): with open(os.path.join(os.path.dirname(__snake_case ) , __snake_case ) ) as input_file: _UpperCamelCase = [ [int(__snake_case ) for element in line.split(''',''' )] f...
10
import os import tempfile import unittest from transformers import FlaubertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, rand...
10
1
from abc import ABC, abstractmethod from typing import List, Optional class lowerCAmelCase_ ( __lowercase ): def __init__( self : Union[str, Any] ): # test for the above condition self.test() def UpperCamelCase_ ( self : Unio...
10
from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTes...
10
1
def _snake_case ( ): return [ a * b * (1000 - a - b) for a in range(1 , 999 ) for b in range(__snake_case , 999 ) if (a * a + b * b == (1000 - a - b) ** 2) ][0] if __name__ == "__main__": print(f'{solution() = }')
10
from __future__ import annotations _lowerCAmelCase = [True] * 1_000_001 _lowerCAmelCase = 2 while i * i <= 1_000_000: if seive[i]: for j in range(i * i, 1_000_001, i): _lowerCAmelCase = False i += 1 def _snake_case ( __snake_case ): return seive[n...
10
1
import os import unittest from transformers.models.cpmant.tokenization_cpmant import VOCAB_FILES_NAMES, CpmAntTokenizer from transformers.testing_utils import require_jieba, tooslow from ...test_tokenization_common import TokenizerTesterMixin @require_jieba class lowerCAmelCase_ ( __lowercase, unittest...
10
import unittest from transformers import DebertaVaTokenizer, DebertaVaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin _lowerCAmelCase = get_tests_dir("fixtures/spiece.model") @requ...
10
1
import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters _lowerCAmelCase = (720, 1_280) # Height, Width _lowerCAmelCase = (0.4, 0.6) # if height or width lower than this scale, drop it. _lowerCAmelCase = 1 / 100 _lowerCAmelCase ...
10
import sys from collections import defaultdict class lowerCAmelCase_ : def __init__( self : Optional[int] ): _UpperCamelCase = [] def UpperCamelCase_ ( self : Any , _A : str ): return self.node_position[verte...
10
1
import random import torch from huggingface_hub import HfApi from diffusers import UNetaDModel _lowerCAmelCase = HfApi() _lowerCAmelCase = {} # fmt: off _lowerCAmelCase = torch.tensor([ -0.7515, -1.6883, 0.2420, 0.0300, 0.6347, 1.3433, -1.1743, -3.7467, 1.2342, -2.2485, 0.4636, 0.8076,...
10
import logging import os from .state import PartialState class lowerCAmelCase_ ( logging.LoggerAdapter ): @staticmethod def UpperCamelCase_ ( _A : Any ): _UpperCamelCase = PartialState() return not main_process_only or (main_proces...
10
1
# NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from ...utils import deprecate from ..controlnet.multicontrolnet import MultiControlNetModel # noqa: F401 from ..controlnet.pipeline_controlnet import ...
10
import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _lowerCAmelCase = "▁" _lowerCAmelCase ...
10
1
import os from math import logaa def _snake_case ( __snake_case = "base_exp.txt" ): _UpperCamelCase = 0 _UpperCamelCase = 0 for i, line in enumerate(open(os.path.join(os.path.dirname(__snake_case ) , __snake_case ) ) ): _UpperCamelCase ...
10
import gc import unittest import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DDPMScheduler, PriorTransformer, StableUnCLIPPipeline, UNetaDConditionModel, ) from diffusers.pipeline...
10
1
import math from collections import defaultdict from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin, SchedulerOutput def _snake_case ( __snake_cas...
10
from __future__ import annotations import math import numpy as np from numpy.linalg import norm def _snake_case ( __snake_case , __snake_case ): return math.sqrt(sum(pow(a - b , 2 ) for a, b in zip(__snake_case , __snake_case ) ) ) def _snake_case ( __snake_cas...
10
1
from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging _lowerCAmelCase = loggin...
10
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow f...
10
1
_lowerCAmelCase = { "a": "AAAAA", "b": "AAAAB", "c": "AAABA", "d": "AAABB", "e": "AABAA", "f": "AABAB", "g": "AABBA", "h": "AABBB", "i": "ABAAA", "j": "BBBAA", "k": "ABAAB", "l": "ABABA", "m": "ABABB", "n": "ABBAA", "o": "ABBAB", "p": "ABBBA", "...
10
import random import torch from huggingface_hub import HfApi from diffusers import UNetaDModel _lowerCAmelCase = HfApi() _lowerCAmelCase = {} # fmt: off _lowerCAmelCase = torch.tensor([ -0.7515, -1.6883, 0.2420, 0.0300, 0.6347, 1.3433, -1.1743, -3.7467, 1.2342, -2.2485, 0.4636, 0.8076,...
10
1
from __future__ import annotations import math _lowerCAmelCase = "2020.9.26" _lowerCAmelCase = "xcodz-dot, cclaus, dhruvmanila" def _snake_case ( __snake_case , __snake_case , __snake_case , __snake_case , __snake_case ): if not all(isinstance(__snake_case , (flo...
10
from typing import List from .keymap import KEYMAP, get_character def _snake_case ( __snake_case ): def decorator(__snake_case ): _UpperCamelCase = getattr(__snake_case , '''handle_key''' , [] ) handle += [key] setattr(__snake_case , ''...
10
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase = logging.get_logger(__name__) _lowerCAmelCase = { "microsoft/trocr-base-handwritten": ( "https://huggingface.co/microsoft/trocr-base-handwritten/resolve/main/config.json" ), # See all T...
10
import unittest from transformers import ( MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TextGenerationPipeline, logging, pipeline, ) from transformers.testing_utils import ( CaptureLogger, is_pipeline_test, require_accelerate, require_tf, require_torch, requi...
10
1
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, TensorType _lowerCAmelCase = logging.get_...
10
def _snake_case ( __snake_case = 100 ): _UpperCamelCase = (n * (n + 1) // 2) ** 2 _UpperCamelCase = n * (n + 1) * (2 * n + 1) // 6 return sum_cubes - sum_squares if __name__ == "__main__": print(f'{solution() = }')
10
1
def _snake_case ( ): return [list(range(1000 - i , -1000 - i , -1 ) ) for i in range(1000 )] _lowerCAmelCase = generate_large_matrix() _lowerCAmelCase = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], ...
10
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDAR...
10
1
from string import ascii_lowercase, ascii_uppercase def _snake_case ( __snake_case ): if not sentence: return "" _UpperCamelCase = dict(zip(__snake_case , __snake_case ) ) return lower_to_upper.get(sentence[0] , sentence[0] ) + sentence[1:] if __...
10
import os import re import shutil import sys import tempfile import unittest import black _lowerCAmelCase = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, "utils")) import check_copies # noqa: E402 # This is the reference code that ...
10
1
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow f...
10
import json import logging import os import re import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import datasets import numpy as np import torch import torchaudio from packaging import version from torch import nn import transformers from transformers import ( ...
10
1
import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class lowerCAmelCase_ ( __lowercase ): UpperCAmelCase = (PNDMScheduler,) UpperCAmelCase = (("num_inference_steps", 50),) def UpperCamelCase_...
10
import math class lowerCAmelCase_ : def __init__( self : Tuple , _A : int=0 ): # a graph with Node 0,1,...,N-1 _UpperCamelCase = n _UpperCamelCase = [ [math.inf for j in range(0 , _A )] for i in range(0 , _A...
10
1
from typing import TYPE_CHECKING from ..utils import _LazyModule _lowerCAmelCase = { "config": [ "EXTERNAL_DATA_FORMAT_SIZE_LIMIT", "OnnxConfig", "OnnxConfigWithPast", "OnnxSeq2SeqConfigWithPast", "PatchingSpec", ], "convert": ["export", "validate_model_outp...
10
import dataclasses import json import warnings from dataclasses import dataclass, field from time import time from typing import List from ..utils import logging _lowerCAmelCase = logging.get_logger(__name__) def _snake_case ( __snake_case=None , __snake_case=None ): return field(default_...
10
1
def _snake_case ( __snake_case , __snake_case ): if a < 0 or b < 0: raise ValueError('''the value of both inputs must be positive''' ) _UpperCamelCase = str(bin(__snake_case ) )[2:] # remove the leading "0b" _UpperCamelCase = str(bin(__snake...
10
import inspect import warnings from typing import Any, Dict, Optional, Union from packaging import version def _snake_case ( *__snake_case , __snake_case = None , __snake_case=True , __snake_case=2 ): from .. import __version__ _UpperCamelCase = take_from _UpperCame...
10
1
from .imports import is_rich_available if is_rich_available(): from rich.traceback import install install(show_locals=False) else: raise ModuleNotFoundError("To use the rich extension, install rich with `pip install rich`")
10
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCo...
10
1
import unittest import numpy as np from transformers import RobertaConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.model...
10
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase = logging.get_logger(__name__) _lowerCAmelCase = { "microsoft/trocr-base-handwritten": ( "https://huggingface.co/microsoft/trocr-base-handwritten/resolve/main/config.json" ), # See all T...
10
1
def _snake_case ( __snake_case , __snake_case ): _UpperCamelCase = len(__snake_case ) print('''The following activities are selected:''' ) # The first activity is always selected _UpperCamelCase = 0 print(__snake_case , end=''',''' ) ...
10
import os import tempfile import unittest from transformers import FlaubertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, rand...
10
1
import unittest from transformers import load_tool from .test_tools_common import ToolTesterMixin class lowerCAmelCase_ ( unittest.TestCase, __lowercase ): def UpperCamelCase_ ( self : Tuple ): _UpperCamelCase = load_tool('''text-classification''' ...
10
from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTes...
10
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCAmelCase = logging.get_logger(__name__) _lowerCAmelCase = { "kssteven/ibert-roberta-base": "https://huggingface.co/k...
10
from __future__ import annotations _lowerCAmelCase = [True] * 1_000_001 _lowerCAmelCase = 2 while i * i <= 1_000_000: if seive[i]: for j in range(i * i, 1_000_001, i): _lowerCAmelCase = False i += 1 def _snake_case ( __snake_case ): return seive[n...
10
1
import json import logging import os import re import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import datasets import numpy as np import torch import torchaudio from packaging import version from torch import nn import transformers from transformers import ( ...
10
import unittest from transformers import DebertaVaTokenizer, DebertaVaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin _lowerCAmelCase = get_tests_dir("fixtures/spiece.model") @requ...
10
1
# flake8: noqa # Lint as: python3 _lowerCAmelCase = [ "VerificationMode", "Version", "disable_progress_bar", "enable_progress_bar", "is_progress_bar_enabled", "experimental", ] from .info_utils import VerificationMode from .logging import disable_progress_bar, enable_progress_bar, is_pro...
10
import sys from collections import defaultdict class lowerCAmelCase_ : def __init__( self : Optional[int] ): _UpperCamelCase = [] def UpperCamelCase_ ( self : Any , _A : str ): return self.node_position[verte...
10
1
from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTes...
10
import logging import os from .state import PartialState class lowerCAmelCase_ ( logging.LoggerAdapter ): @staticmethod def UpperCamelCase_ ( _A : Any ): _UpperCamelCase = PartialState() return not main_process_only or (main_proces...
10
1
import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFI...
10
import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _lowerCAmelCase = "▁" _lowerCAmelCase ...
10
1
from __future__ import annotations from typing import TypedDict class lowerCAmelCase_ ( __lowercase ): UpperCAmelCase = 42 UpperCAmelCase = 42 def _snake_case ( __snake_case ): if not isinstance(__snake_case , __snake_case ): raise...
10
import gc import unittest import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DDPMScheduler, PriorTransformer, StableUnCLIPPipeline, UNetaDConditionModel, ) from diffusers.pipeline...
10
1
from __future__ import annotations from collections import deque from collections.abc import Sequence from dataclasses import dataclass from typing import Any @dataclass class lowerCAmelCase_ : UpperCAmelCase = 42 UpperCAmelCase = None UpperCAmelCase = None ...
10
from __future__ import annotations import math import numpy as np from numpy.linalg import norm def _snake_case ( __snake_case , __snake_case ): return math.sqrt(sum(pow(a - b , 2 ) for a, b in zip(__snake_case , __snake_case ) ) ) def _snake_case ( __snake_cas...
10
1
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device _lowerCAmelCase = False class lowerCAmelCase_ ( unittest.TestCase ): p...
10
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow f...
10
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _lowerCAmelCase = { "configuration_efficientformer": [ "EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "Effi...
10
import random import torch from huggingface_hub import HfApi from diffusers import UNetaDModel _lowerCAmelCase = HfApi() _lowerCAmelCase = {} # fmt: off _lowerCAmelCase = torch.tensor([ -0.7515, -1.6883, 0.2420, 0.0300, 0.6347, 1.3433, -1.1743, -3.7467, 1.2342, -2.2485, 0.4636, 0.8076,...
10
1
from functools import lru_cache def _snake_case ( __snake_case ): _UpperCamelCase = 2 _UpperCamelCase = set() while i * i <= n: if n % i: i += 1 else: n //= i factors.add(__snake_case ) if n > 1: ...
10
from typing import List from .keymap import KEYMAP, get_character def _snake_case ( __snake_case ): def decorator(__snake_case ): _UpperCamelCase = getattr(__snake_case , '''handle_key''' , [] ) handle += [key] setattr(__snake_case , ''...
10
1
def _snake_case ( __snake_case ): return str(__snake_case ) == str(__snake_case )[::-1] def _snake_case ( __snake_case ): return int(__snake_case ) + int(str(__snake_case )[::-1] ) def _snake_case ( __snake_case = 10000 ): _UpperCamelCase =...
10
import unittest from transformers import ( MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TextGenerationPipeline, logging, pipeline, ) from transformers.testing_utils import ( CaptureLogger, is_pipeline_test, require_accelerate, require_tf, require_torch, requi...
10
1
_lowerCAmelCase = "Input must be a string of 8 numbers plus letter" _lowerCAmelCase = "TRWAGMYFPDXBNJZSQVHLCKE" def _snake_case ( __snake_case ): if not isinstance(__snake_case , __snake_case ): _UpperCamelCase = f"""Expected string as input, found {type(__s...
10
def _snake_case ( __snake_case = 100 ): _UpperCamelCase = (n * (n + 1) // 2) ** 2 _UpperCamelCase = n * (n + 1) * (2 * n + 1) // 6 return sum_cubes - sum_squares if __name__ == "__main__": print(f'{solution() = }')
10
1
import os def _snake_case ( __snake_case = "matrix.txt" ): with open(os.path.join(os.path.dirname(__snake_case ) , __snake_case ) ) as in_file: _UpperCamelCase = in_file.read() _UpperCamelCase = [[int(__snake_case ) for cell in row.split('''...
10
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDAR...
10
1
from collections.abc import Iterable from typing import Generic, TypeVar _lowerCAmelCase = TypeVar("_T") class lowerCAmelCase_ ( Generic[_T] ): def __init__( self : str , _A : Iterable[_T] | None = None ): _UpperCamelCase = list(iterable ...
10
import os import re import shutil import sys import tempfile import unittest import black _lowerCAmelCase = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, "utils")) import check_copies # noqa: E402 # This is the reference code that ...
10
1
import sys from collections import defaultdict class lowerCAmelCase_ : def __init__( self : Optional[int] ): _UpperCamelCase = [] def UpperCamelCase_ ( self : Any , _A : str ): return self.node_position[verte...
10
import json import logging import os import re import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import datasets import numpy as np import torch import torchaudio from packaging import version from torch import nn import transformers from transformers import ( ...
10
1
def _snake_case ( __snake_case , __snake_case ): # "extended trapezoidal rule" # int(f) = dx/2 * (f1 + 2f2 + ... + fn) _UpperCamelCase = (boundary[1] - boundary[0]) / steps _UpperCamelCase = boundary[0] _UpperCamelCase = boundary[1] _UpperCa...
10
import math class lowerCAmelCase_ : def __init__( self : Tuple , _A : int=0 ): # a graph with Node 0,1,...,N-1 _UpperCamelCase = n _UpperCamelCase = [ [math.inf for j in range(0 , _A )] for i in range(0 , _A...
10
1
from datetime import datetime import matplotlib.pyplot as plt import torch def _snake_case ( __snake_case ): for param in module.parameters(): _UpperCamelCase = False def _snake_case ( ): _UpperCamelCase = '''cuda''' if torch.cuda.is_available() else '''cpu'''...
10
import dataclasses import json import warnings from dataclasses import dataclass, field from time import time from typing import List from ..utils import logging _lowerCAmelCase = logging.get_logger(__name__) def _snake_case ( __snake_case=None , __snake_case=None ): return field(default_...
10
1
from __future__ import annotations def _snake_case ( __snake_case ): if len(__snake_case ) == 0: return array _UpperCamelCase , _UpperCamelCase = min(__snake_case ), max(__snake_case ) # Compute the variables _UpperCamelCase = _max -...
10
import inspect import warnings from typing import Any, Dict, Optional, Union from packaging import version def _snake_case ( *__snake_case , __snake_case = None , __snake_case=True , __snake_case=2 ): from .. import __version__ _UpperCamelCase = take_from _UpperCame...
10
1
import gc import unittest import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DDPMScheduler, PriorTransformer, StableUnCLIPPipeline, UNetaDConditionModel, ) from diffusers.pipeline...
10
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCo...
10
1
import argparse import fairseq import torch from torch import nn from transformers import ( MBartaaTokenizer, MBartConfig, MBartForCausalLM, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verb...
10
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase = logging.get_logger(__name__) _lowerCAmelCase = { "microsoft/trocr-base-handwritten": ( "https://huggingface.co/microsoft/trocr-base-handwritten/resolve/main/config.json" ), # See all T...
10
1
import importlib import os import sys # This is required to make the module import works (when the python process is running from the root of the repo) sys.path.append(".") def _snake_case ( __snake_case ): _UpperCamelCase = test_file.split(os.path.sep ) if components[0:2] != ["te...
10
import os import tempfile import unittest from transformers import FlaubertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, rand...
10
1
import torch from diffusers import UnCLIPScheduler from .test_schedulers import SchedulerCommonTest class lowerCAmelCase_ ( __lowercase ): UpperCAmelCase = (UnCLIPScheduler,) def UpperCamelCase_ ( self : Any , **_A : int ): _UpperCa...
10
from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTes...
10
1
import json import os from typing import Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _lowerCAmelCase = logging.get_logger(__name__) _lowerCAmelCase = { "vocab_file": "vocab.json", "merges_file": "merges.txt", } _lowerC...
10
from __future__ import annotations _lowerCAmelCase = [True] * 1_000_001 _lowerCAmelCase = 2 while i * i <= 1_000_000: if seive[i]: for j in range(i * i, 1_000_001, i): _lowerCAmelCase = False i += 1 def _snake_case ( __snake_case ): return seive[n...
10
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCAmelCase = { "configuration_swinv2": ["SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP", "Swinv2Config"], } try: if not is_torch_available(): raise OptionalDependencyNotAvaila...
10
import unittest from transformers import DebertaVaTokenizer, DebertaVaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin _lowerCAmelCase = get_tests_dir("fixtures/spiece.model") @requ...
10
1
import logging import math import os from dataclasses import dataclass, field from glob import glob from typing import Optional from torch.utils.data import ConcatDataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_WITH_LM_HEAD_MAPPING, AutoConfig, AutoModelWithLMHead, Aut...
10
import sys from collections import defaultdict class lowerCAmelCase_ : def __init__( self : Optional[int] ): _UpperCamelCase = [] def UpperCamelCase_ ( self : Any , _A : str ): return self.node_position[verte...
10
1
from dataclasses import dataclass from typing import Dict, Optional, Union import torch import torch.nn.functional as F from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .attention_processor import At...
10
import logging import os from .state import PartialState class lowerCAmelCase_ ( logging.LoggerAdapter ): @staticmethod def UpperCamelCase_ ( _A : Any ): _UpperCamelCase = PartialState() return not main_process_only or (main_proces...
10
1
import logging from dataclasses import dataclass, field from pathlib import Path from typing import Optional, Union from .generation.configuration_utils import GenerationConfig from .training_args import TrainingArguments from .utils import add_start_docstrings _lowerCAmelCase = logging.getLogger(__name__) @d...
10
import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _lowerCAmelCase = "▁" _lowerCAmelCase ...
10
1
def _snake_case ( __snake_case , __snake_case , __snake_case , __snake_case , __snake_case ): if index == number_of_items: return 0 _UpperCamelCase = 0 _UpperCamelCase = 0 _UpperCamelCase = knapsack(__snake_case , __snake...
10
import gc import unittest import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DDPMScheduler, PriorTransformer, StableUnCLIPPipeline, UNetaDConditionModel, ) from diffusers.pipeline...
10
1
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCAmelCase = {"configuration_focalnet": ["FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP", "FocalNetConfig"]} try: if not is_torch_available(): ...
10
from __future__ import annotations import math import numpy as np from numpy.linalg import norm def _snake_case ( __snake_case , __snake_case ): return math.sqrt(sum(pow(a - b , 2 ) for a, b in zip(__snake_case , __snake_case ) ) ) def _snake_case ( __snake_cas...
10
1
def _snake_case ( __snake_case ): _UpperCamelCase = len(__snake_case ) _UpperCamelCase = sum(__snake_case ) _UpperCamelCase = [[False for x in range(s + 1 )] for y in range(n + 1 )] for i in range(1 , n + 1 ): _Upp...
10
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow f...
10
1
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel from diffusers import DDIMScheduler, LDMPipeline, UNetaDModel, VQModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class lower...
10
import random import torch from huggingface_hub import HfApi from diffusers import UNetaDModel _lowerCAmelCase = HfApi() _lowerCAmelCase = {} # fmt: off _lowerCAmelCase = torch.tensor([ -0.7515, -1.6883, 0.2420, 0.0300, 0.6347, 1.3433, -1.1743, -3.7467, 1.2342, -2.2485, 0.4636, 0.8076,...
10
1
import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() _lowerCAmelCase = logging.get_logger("transformers.models.speecht5") def _snake_case ( __snake_case , __snake_case , __snake_case ): ...
10
from typing import List from .keymap import KEYMAP, get_character def _snake_case ( __snake_case ): def decorator(__snake_case ): _UpperCamelCase = getattr(__snake_case , '''handle_key''' , [] ) handle += [key] setattr(__snake_case , ''...
10
1
from collections.abc import Callable class lowerCAmelCase_ : def __init__( self : Tuple , _A : Callable | None = None ): # Stores actual heap items. _UpperCamelCase = [] # Stores indexes of each item for supporting updates and d...
10
import unittest from transformers import ( MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TextGenerationPipeline, logging, pipeline, ) from transformers.testing_utils import ( CaptureLogger, is_pipeline_test, require_accelerate, require_tf, require_torch, requi...
10
1
def _snake_case ( ): _UpperCamelCase = 0 for i in range(1 , 1001 ): total += i**i return str(__snake_case )[-10:] if __name__ == "__main__": print(solution())
10
def _snake_case ( __snake_case = 100 ): _UpperCamelCase = (n * (n + 1) // 2) ** 2 _UpperCamelCase = n * (n + 1) * (2 * n + 1) // 6 return sum_cubes - sum_squares if __name__ == "__main__": print(f'{solution() = }')
10
1
def _snake_case ( __snake_case ): _UpperCamelCase = int(__snake_case ) if decimal in (0, 1): # Exit cases for the recursion return str(__snake_case ) _UpperCamelCase , _UpperCamelCase = divmod(__snake_case , 2 ) return binary_rec...
10
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDAR...
10
1
from __future__ import annotations _lowerCAmelCase = [] def _snake_case ( __snake_case , __snake_case , __snake_case ): for i in range(len(__snake_case ) ): if board[row][i] == 1: return False for i in range(len(__snake_case ) ): ...
10
import os import re import shutil import sys import tempfile import unittest import black _lowerCAmelCase = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, "utils")) import check_copies # noqa: E402 # This is the reference code that ...
10
1
from __future__ import annotations def _snake_case ( __snake_case , __snake_case = None ): _UpperCamelCase = word_bank or [] # create a table _UpperCamelCase = len(__snake_case ) + 1 _UpperCamelCase = [] for _ in range(__snake_case ): ...
10
import json import logging import os import re import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import datasets import numpy as np import torch import torchaudio from packaging import version from torch import nn import transformers from transformers import ( ...
10
1
def _snake_case ( __snake_case , __snake_case ): if discount_rate < 0: raise ValueError('''Discount rate cannot be negative''' ) if not cash_flows: raise ValueError('''Cash flows list cannot be empty''' ) _UpperCamelCase = sum( cash_flow /...
10
import math class lowerCAmelCase_ : def __init__( self : Tuple , _A : int=0 ): # a graph with Node 0,1,...,N-1 _UpperCamelCase = n _UpperCamelCase = [ [math.inf for j in range(0 , _A )] for i in range(0 , _A...
10
1
from ....configuration_utils import PretrainedConfig from ....utils import logging _lowerCAmelCase = logging.get_logger(__name__) # TODO: upload to AWS _lowerCAmelCase = { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/config.json" ...
10
import dataclasses import json import warnings from dataclasses import dataclass, field from time import time from typing import List from ..utils import logging _lowerCAmelCase = logging.get_logger(__name__) def _snake_case ( __snake_case=None , __snake_case=None ): return field(default_...
10
1
from __future__ import annotations def _snake_case ( __snake_case ): return [ord(__snake_case ) - 96 for elem in plain] def _snake_case ( __snake_case ): return "".join(chr(elem + 96 ) for elem in encoded ) def _snake_case ( ): _UpperCamelCase = encod...
10
import inspect import warnings from typing import Any, Dict, Optional, Union from packaging import version def _snake_case ( *__snake_case , __snake_case = None , __snake_case=True , __snake_case=2 ): from .. import __version__ _UpperCamelCase = take_from _UpperCame...
10
1