code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
import os import zipfile import pytest from datasets.utils.extract import ( BzipaExtractor, Extractor, GzipExtractor, LzaExtractor, SevenZipExtractor, TarExtractor, XzExtractor, ZipExtractor, ZstdExtractor, ) from .utils import require_lza, require_pyazr, require_zstandard @pytest.mark.parametrize( 'compression_format, is_archive' , [ ('7z', True), ('bz2', False), ('gzip', False), ('lz4', False), ('tar', True), ('xz', False), ('zip', True), ('zstd', False), ] , ) def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , ) -> Dict: _lowercase : Optional[Any] = { '7z': (seven_zip_file, SevenZipExtractor), 'bz2': (bza_file, BzipaExtractor), 'gzip': (gz_file, GzipExtractor), 'lz4': (lza_file, LzaExtractor), 'tar': (tar_file, TarExtractor), 'xz': (xz_file, XzExtractor), 'zip': (zip_file, ZipExtractor), 'zstd': (zstd_file, ZstdExtractor), } _lowercase , _lowercase : Any = input_paths_and_base_extractors[compression_format] if input_path is None: _lowercase : List[str] = F"""for '{compression_format}' compression_format, """ if compression_format == "7z": reason += require_pyazr.kwargs["reason"] elif compression_format == "lz4": reason += require_lza.kwargs["reason"] elif compression_format == "zstd": reason += require_zstandard.kwargs["reason"] pytest.skip(SCREAMING_SNAKE_CASE ) assert base_extractor.is_extractable(SCREAMING_SNAKE_CASE ) _lowercase : Any = tmp_path / ('extracted' if is_archive else 'extracted.txt') base_extractor.extract(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if is_archive: assert output_path.is_dir() for file_path in output_path.iterdir(): assert file_path.name == text_file.name _lowercase : Optional[int] = file_path.read_text(encoding='utf-8' ) else: _lowercase : Any = output_path.read_text(encoding='utf-8' ) _lowercase : str = text_file.read_text(encoding='utf-8' ) assert extracted_file_content == expected_file_content @pytest.mark.parametrize( 'compression_format, is_archive' , [ ('7z', True), ('bz2', False), ('gzip', False), ('lz4', False), ('tar', True), ('xz', False), ('zip', True), ('zstd', False), ] , ) def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , ) -> Optional[Any]: _lowercase : Tuple = { '7z': seven_zip_file, 'bz2': bza_file, 'gzip': gz_file, 'lz4': lza_file, 'tar': tar_file, 'xz': xz_file, 'zip': zip_file, 'zstd': zstd_file, } _lowercase : Any = input_paths[compression_format] if input_path is None: _lowercase : List[str] = F"""for '{compression_format}' compression_format, """ if compression_format == "7z": reason += require_pyazr.kwargs["reason"] elif compression_format == "lz4": reason += require_lza.kwargs["reason"] elif compression_format == "zstd": reason += require_zstandard.kwargs["reason"] pytest.skip(SCREAMING_SNAKE_CASE ) _lowercase : Tuple = Extractor.infer_extractor_format(SCREAMING_SNAKE_CASE ) assert extractor_format is not None _lowercase : int = tmp_path / ('extracted' if is_archive else 'extracted.txt') Extractor.extract(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if is_archive: assert output_path.is_dir() for file_path in output_path.iterdir(): assert file_path.name == text_file.name _lowercase : List[Any] = file_path.read_text(encoding='utf-8' ) else: _lowercase : List[str] = output_path.read_text(encoding='utf-8' ) _lowercase : str = text_file.read_text(encoding='utf-8' ) assert extracted_file_content == expected_file_content @pytest.fixture def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Union[str, Any]: import tarfile _lowercase : List[Any] = tmp_path / 'data_dot_dot' directory.mkdir() _lowercase : Optional[Any] = directory / 'tar_file_with_dot_dot.tar' with tarfile.TarFile(SCREAMING_SNAKE_CASE , 'w' ) as f: f.add(SCREAMING_SNAKE_CASE , arcname=os.path.join('..' , text_file.name ) ) return path @pytest.fixture def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> Any: import tarfile _lowercase : List[Any] = tmp_path / 'data_sym_link' directory.mkdir() _lowercase : str = directory / 'tar_file_with_sym_link.tar' os.symlink('..' , directory / 'subdir' , target_is_directory=SCREAMING_SNAKE_CASE ) with tarfile.TarFile(SCREAMING_SNAKE_CASE , 'w' ) as f: f.add(str(directory / 'subdir' ) , arcname='subdir' ) # str required by os.readlink on Windows and Python < 3.8 return path @pytest.mark.parametrize( 'insecure_tar_file, error_log' , [('tar_file_with_dot_dot', 'illegal path'), ('tar_file_with_sym_link', 'Symlink')] , ) def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Union[str, Any]: _lowercase : Optional[int] = { 'tar_file_with_dot_dot': tar_file_with_dot_dot, 'tar_file_with_sym_link': tar_file_with_sym_link, } _lowercase : str = insecure_tar_files[insecure_tar_file] _lowercase : List[Any] = tmp_path / 'extracted' TarExtractor.extract(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) assert caplog.text for record in caplog.records: assert record.levelname == "ERROR" assert error_log in record.msg def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> Optional[int]: # We should have less false positives than zipfile.is_zipfile # We do that by checking only the magic number _lowercase : List[str] = tmpdir / 'not_a_zip_file' # From: https://github.com/python/cpython/pull/5053 _lowercase : List[str] = ( b'\x89PNG\r\n\x1a\n\x00\x00\x00\rIHDR\x00\x00\x00\x01\x00\x00' b'\x00\x02\x08\x06\x00\x00\x00\x99\x81\xb6\'\x00\x00\x00\x15I' b'DATx\x01\x01\n\x00\xf5\xff\x00PK\x05\x06\x00PK\x06\x06\x07' b'\xac\x01N\xc6|a\r\x00\x00\x00\x00IEND\xaeB`\x82' ) with not_a_zip_file.open('wb' ) as f: f.write(SCREAMING_SNAKE_CASE ) assert zipfile.is_zipfile(str(SCREAMING_SNAKE_CASE ) ) # is a false positive for `zipfile` assert not ZipExtractor.is_extractable(SCREAMING_SNAKE_CASE ) # but we're right
66
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto.configuration_auto import CONFIG_MAPPING UpperCamelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : Any = "upernet" def __init__( self , _lowerCAmelCase=None , _lowerCAmelCase=5_1_2 , _lowerCAmelCase=0.02 , _lowerCAmelCase=[1, 2, 3, 6] , _lowerCAmelCase=True , _lowerCAmelCase=0.4 , _lowerCAmelCase=3_8_4 , _lowerCAmelCase=2_5_6 , _lowerCAmelCase=1 , _lowerCAmelCase=False , _lowerCAmelCase=2_5_5 , **_lowerCAmelCase , ): super().__init__(**_lowerCAmelCase ) if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.' ) _lowercase : Optional[Any] = CONFIG_MAPPING['resnet'](out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): _lowercase : List[Any] = backbone_config.get('model_type' ) _lowercase : str = CONFIG_MAPPING[backbone_model_type] _lowercase : Tuple = config_class.from_dict(_lowerCAmelCase ) _lowercase : Optional[Any] = backbone_config _lowercase : Any = hidden_size _lowercase : Any = initializer_range _lowercase : Tuple = pool_scales _lowercase : List[Any] = use_auxiliary_head _lowercase : Optional[Any] = auxiliary_loss_weight _lowercase : Any = auxiliary_in_channels _lowercase : Any = auxiliary_channels _lowercase : List[str] = auxiliary_num_convs _lowercase : List[str] = auxiliary_concat_input _lowercase : Tuple = loss_ignore_index def __a ( self ): _lowercase : str = copy.deepcopy(self.__dict__ ) _lowercase : Tuple = self.backbone_config.to_dict() _lowercase : int = self.__class__.model_type return output
66
1
import unittest import numpy as np from transformers import BertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.bert.modeling_flax_bert import ( FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, ) class lowerCAmelCase_ ( unittest.TestCase ): def __init__( self , _lowerCAmelCase , _lowerCAmelCase=1_3 , _lowerCAmelCase=7 , _lowerCAmelCase=True , _lowerCAmelCase=True , _lowerCAmelCase=True , _lowerCAmelCase=True , _lowerCAmelCase=9_9 , _lowerCAmelCase=3_2 , _lowerCAmelCase=5 , _lowerCAmelCase=4 , _lowerCAmelCase=3_7 , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.1 , _lowerCAmelCase=5_1_2 , _lowerCAmelCase=1_6 , _lowerCAmelCase=2 , _lowerCAmelCase=0.02 , _lowerCAmelCase=4 , ): _lowercase : Optional[Any] = parent _lowercase : Any = batch_size _lowercase : str = seq_length _lowercase : Union[str, Any] = is_training _lowercase : Tuple = use_attention_mask _lowercase : List[str] = use_token_type_ids _lowercase : Tuple = use_labels _lowercase : Tuple = vocab_size _lowercase : List[Any] = hidden_size _lowercase : List[Any] = num_hidden_layers _lowercase : List[Any] = num_attention_heads _lowercase : Any = intermediate_size _lowercase : List[str] = hidden_act _lowercase : Dict = hidden_dropout_prob _lowercase : int = attention_probs_dropout_prob _lowercase : str = max_position_embeddings _lowercase : Tuple = type_vocab_size _lowercase : Optional[int] = type_sequence_label_size _lowercase : Optional[int] = initializer_range _lowercase : List[Any] = num_choices def __a ( self ): _lowercase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowercase : str = None if self.use_attention_mask: _lowercase : int = random_attention_mask([self.batch_size, self.seq_length] ) _lowercase : Optional[Any] = None if self.use_token_type_ids: _lowercase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowercase : Any = BertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def __a ( self ): _lowercase : List[Any] = self.prepare_config_and_inputs() _lowercase , _lowercase , _lowercase , _lowercase : Dict = config_and_inputs _lowercase : Dict = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': attention_mask} return config, inputs_dict def __a ( self ): _lowercase : Tuple = self.prepare_config_and_inputs() _lowercase , _lowercase , _lowercase , _lowercase : List[Any] = config_and_inputs _lowercase : Any = True _lowercase : str = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) _lowercase : Tuple = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, attention_mask, encoder_hidden_states, encoder_attention_mask, ) @require_flax class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : Optional[Any] = True _UpperCamelCase : int = ( ( FlaxBertModel, FlaxBertForPreTraining, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForQuestionAnswering, FlaxBertForNextSentencePrediction, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertForQuestionAnswering, ) if is_flax_available() else () ) def __a ( self ): _lowercase : Optional[Any] = FlaxBertModelTester(self ) @slow def __a ( self ): # Only check this for base model, not necessary for all model classes. # This will also help speed-up tests. _lowercase : Union[str, Any] = FlaxBertModel.from_pretrained('bert-base-cased' ) _lowercase : int = model(np.ones((1, 1) ) ) self.assertIsNotNone(_lowerCAmelCase )
66
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: if a < 0 or b < 0: raise ValueError('the value of both inputs must be positive' ) _lowercase : str = str(bin(SCREAMING_SNAKE_CASE ) )[2:] # remove the leading "0b" _lowercase : int = str(bin(SCREAMING_SNAKE_CASE ) )[2:] # remove the leading "0b" _lowercase : List[Any] = max(len(SCREAMING_SNAKE_CASE ) , len(SCREAMING_SNAKE_CASE ) ) return "0b" + "".join( str(int(char_a == '1' and char_b == '1' ) ) for char_a, char_b in zip(a_binary.zfill(SCREAMING_SNAKE_CASE ) , b_binary.zfill(SCREAMING_SNAKE_CASE ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
66
1
def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : Optional[Any] = 1 for i in range(1 , num + 1 ): fact *= i return fact def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : Union[str, Any] = 0 while number > 0: _lowercase : List[Any] = number % 10 sum_of_digits += last_digit _lowercase : Tuple = number // 10 # Removing the last_digit from the given number return sum_of_digits def __magic_name__ ( SCREAMING_SNAKE_CASE = 100 ) -> int: _lowercase : List[Any] = factorial(SCREAMING_SNAKE_CASE ) _lowercase : Optional[Any] = split_and_add(SCREAMING_SNAKE_CASE ) return result if __name__ == "__main__": print(solution(int(input("Enter the Number: ").strip())))
66
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class lowerCAmelCase_ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCamelCase : int = IFInpaintingSuperResolutionPipeline _UpperCamelCase : Any = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"width", "height"} _UpperCamelCase : Union[str, Any] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"original_image"} ) _UpperCamelCase : Tuple = PipelineTesterMixin.required_optional_params - {"latents"} def __a ( self ): return self._get_superresolution_dummy_components() def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : int = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Union[str, Any] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : Union[str, Any] = floats_tensor((1, 3, 1_6, 1_6) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : Optional[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'original_image': original_image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def __a ( self ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def __a ( self ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != 'cuda' , reason='float16 requires CUDA' ) def __a ( self ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __a ( self ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __a ( self ): self._test_save_load_local() def __a ( self ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
66
1
import re from flax.core.frozen_dict import freeze from flax.traverse_util import flatten_dict, unflatten_dict from jax.experimental import PartitionSpec as P # Sentinels UpperCamelCase = object() # For specifying empty leaf dict `{}` UpperCamelCase = object() def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> List[str]: _lowercase : Optional[int] = tuple((re.compile(x + '$' ) for x in qs) ) for i in range(len(SCREAMING_SNAKE_CASE ) - len(SCREAMING_SNAKE_CASE ) + 1 ): _lowercase : Tuple = [x.match(SCREAMING_SNAKE_CASE ) for x, y in zip(SCREAMING_SNAKE_CASE , ks[i:] )] if matches and all(SCREAMING_SNAKE_CASE ): return True return False def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> Dict: def replace(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): for rule, replacement in rules: if _match(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): return replacement return val return replace def __magic_name__ ( ) -> Dict: return [ # embeddings (("transformer", "wpe", "embedding"), P('mp' , SCREAMING_SNAKE_CASE )), (("transformer", "wte", "embedding"), P('mp' , SCREAMING_SNAKE_CASE )), # atention (("attention", "(q_proj|k_proj|v_proj)", "kernel"), P(SCREAMING_SNAKE_CASE , 'mp' )), (("attention", "out_proj", "kernel"), P('mp' , SCREAMING_SNAKE_CASE )), (("attention", "out_proj", "bias"), None), # mlp (("mlp", "c_fc", "kernel"), P(SCREAMING_SNAKE_CASE , 'mp' )), (("mlp", "c_fc", "bias"), P('mp' )), (("mlp", "c_proj", "kernel"), P('mp' , SCREAMING_SNAKE_CASE )), (("mlp", "c_proj", "bias"), None), # layer norms ((r"ln_\d+", "bias"), None), ((r"\d+", r"ln_\d+", "scale"), None), (("ln_f", "bias"), None), (("ln_f", "scale"), None), ] def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> Dict: _lowercase : Optional[int] = _get_partition_rules() _lowercase : Optional[int] = _replacement_rules(SCREAMING_SNAKE_CASE ) _lowercase : Optional[Any] = {k: _unmatched for k in flatten_dict(SCREAMING_SNAKE_CASE )} _lowercase : int = {k: replace(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for k, v in initd.items()} assert _unmatched not in result.values(), "Incomplete partition spec." return freeze(unflatten_dict(SCREAMING_SNAKE_CASE ) )
66
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> tuple[float, float]: # Check if the input is valid if not len(SCREAMING_SNAKE_CASE ) == len(SCREAMING_SNAKE_CASE ) == 3: raise ValueError('Please enter a valid equation.' ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError('Both a & b of two equations can\'t be zero.' ) # Extract the coefficients _lowercase , _lowercase , _lowercase : Tuple = equationa _lowercase , _lowercase , _lowercase : Dict = equationa # Calculate the determinants of the matrices _lowercase : str = aa * ba - aa * ba _lowercase : Any = ca * ba - ca * ba _lowercase : Optional[int] = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError('Infinite solutions. (Consistent system)' ) else: raise ValueError('No solution. (Inconsistent system)' ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: _lowercase : Union[str, Any] = determinant_x / determinant _lowercase : Tuple = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
66
1
import heapq as hq import math from collections.abc import Iterator class lowerCAmelCase_ : def __init__( self , _lowerCAmelCase ): _lowercase : List[str] = str(id_ ) _lowercase : Tuple = None _lowercase : Optional[Any] = None _lowercase : List[Any] = [] _lowercase : List[str] = {} # {vertex:distance} def __lt__( self , _lowerCAmelCase ): return self.key < other.key def __repr__( self ): return self.id def __a ( self , _lowerCAmelCase ): self.neighbors.append(_lowerCAmelCase ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase ): _lowercase : Union[str, Any] = weight def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> List[Any]: # add the neighbors: graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1] , SCREAMING_SNAKE_CASE ) graph[b - 1].add_edge(graph[a - 1] , SCREAMING_SNAKE_CASE ) def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> list: _lowercase : str = [] for u in graph: _lowercase : Optional[Any] = math.inf _lowercase : List[str] = None _lowercase : str = 0 _lowercase : Any = graph[:] while q: _lowercase : str = min(SCREAMING_SNAKE_CASE ) q.remove(SCREAMING_SNAKE_CASE ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): _lowercase : Any = u _lowercase : Dict = u.edges[v.id] for i in range(1 , len(SCREAMING_SNAKE_CASE ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Iterator[tuple]: for u in graph: _lowercase : Optional[Any] = math.inf _lowercase : Optional[int] = None _lowercase : Tuple = 0 _lowercase : Optional[int] = list(SCREAMING_SNAKE_CASE ) hq.heapify(SCREAMING_SNAKE_CASE ) while h: _lowercase : List[Any] = hq.heappop(SCREAMING_SNAKE_CASE ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): _lowercase : Any = u _lowercase : List[Any] = u.edges[v.id] hq.heapify(SCREAMING_SNAKE_CASE ) for i in range(1 , len(SCREAMING_SNAKE_CASE ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def __magic_name__ ( ) -> None: pass if __name__ == "__main__": import doctest doctest.testmod()
66
def __magic_name__ ( SCREAMING_SNAKE_CASE = 50 ) -> int: _lowercase : Optional[int] = [[0] * 3 for _ in range(length + 1 )] for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): different_colour_ways_number[row_length][tile_length - 2] += ( different_colour_ways_number[row_length - tile_start - tile_length][ tile_length - 2 ] + 1 ) return sum(different_colour_ways_number[length] ) if __name__ == "__main__": print(f'''{solution() = }''')
66
1
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyImgaImgPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.text_encoder import MCLIPConfig, MultilingualCLIP from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : List[str] = KandinskyImgaImgPipeline _UpperCamelCase : List[str] = ["prompt", "image_embeds", "negative_image_embeds", "image"] _UpperCamelCase : Dict = [ "prompt", "negative_prompt", "image_embeds", "negative_image_embeds", "image", ] _UpperCamelCase : Optional[int] = [ "generator", "height", "width", "strength", "guidance_scale", "negative_prompt", "num_inference_steps", "return_dict", "guidance_scale", "num_images_per_prompt", "output_type", "return_dict", ] _UpperCamelCase : Tuple = False @property def __a ( self ): return 3_2 @property def __a ( self ): return 3_2 @property def __a ( self ): return self.time_input_dim @property def __a ( self ): return self.time_input_dim * 4 @property def __a ( self ): return 1_0_0 @property def __a ( self ): _lowercase : int = XLMRobertaTokenizerFast.from_pretrained('YiYiXu/tiny-random-mclip-base' ) return tokenizer @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Tuple = MCLIPConfig( numDims=self.cross_attention_dim , transformerDimensions=self.text_embedder_hidden_size , hidden_size=self.text_embedder_hidden_size , intermediate_size=3_7 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=1_0_0_5 , ) _lowercase : Any = MultilingualCLIP(_lowerCAmelCase ) _lowercase : List[str] = text_encoder.eval() return text_encoder @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Tuple = { 'in_channels': 4, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'text_image', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'text_image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } _lowercase : Tuple = UNetaDConditionModel(**_lowerCAmelCase ) return model @property def __a ( self ): return { "block_out_channels": [3_2, 6_4], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 1_2, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[Any] = VQModel(**self.dummy_movq_kwargs ) return model def __a ( self ): _lowercase : Dict = self.dummy_text_encoder _lowercase : Dict = self.dummy_tokenizer _lowercase : Any = self.dummy_unet _lowercase : Optional[int] = self.dummy_movq _lowercase : Tuple = { 'num_train_timesteps': 1_0_0_0, 'beta_schedule': 'linear', 'beta_start': 0.0_00_85, 'beta_end': 0.0_12, 'clip_sample': False, 'set_alpha_to_one': False, 'steps_offset': 0, 'prediction_type': 'epsilon', 'thresholding': False, } _lowercase : int = DDIMScheduler(**_lowerCAmelCase ) _lowercase : Tuple = { 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): _lowercase : str = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : Union[str, Any] = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(seed + 1 ) ).to(_lowerCAmelCase ) # create init_image _lowercase : Tuple = floats_tensor((1, 3, 6_4, 6_4) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : Union[str, Any] = image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowercase : Union[str, Any] = Image.fromarray(np.uinta(_lowerCAmelCase ) ).convert('RGB' ).resize((2_5_6, 2_5_6) ) if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : str = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Optional[int] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : Optional[int] = { 'prompt': 'horse', 'image': init_image, 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'generator': generator, 'height': 6_4, 'width': 6_4, 'num_inference_steps': 1_0, 'guidance_scale': 7.0, 'strength': 0.2, 'output_type': 'np', } return inputs def __a ( self ): _lowercase : List[Any] = 'cpu' _lowercase : Union[str, Any] = self.get_dummy_components() _lowercase : str = self.pipeline_class(**_lowerCAmelCase ) _lowercase : List[str] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Dict = pipe(**self.get_dummy_inputs(_lowerCAmelCase ) ) _lowercase : Tuple = output.images _lowercase : Union[str, Any] = pipe( **self.get_dummy_inputs(_lowerCAmelCase ) , return_dict=_lowerCAmelCase , )[0] _lowercase : Optional[int] = image[0, -3:, -3:, -1] _lowercase : Any = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) _lowercase : Union[str, Any] = np.array( [0.61_47_49_43, 0.6_07_35_39, 0.43_30_85_44, 0.5_92_82_69, 0.47_49_35_95, 0.46_75_59_73, 0.4_61_38_38, 0.45_36_87_97, 0.50_11_92_33] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), F""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), F""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" @slow @require_torch_gpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ): _lowercase : Any = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/kandinsky_img2img_frog.npy' ) _lowercase : Optional[Any] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/cat.png' ) _lowercase : Optional[int] = 'A red cartoon frog, 4k' _lowercase : Optional[Any] = KandinskyPriorPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-1-prior' , torch_dtype=torch.floataa ) pipe_prior.to(_lowerCAmelCase ) _lowercase : Tuple = KandinskyImgaImgPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-1' , torch_dtype=torch.floataa ) _lowercase : Tuple = pipeline.to(_lowerCAmelCase ) pipeline.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Tuple = torch.Generator(device='cpu' ).manual_seed(0 ) _lowercase , _lowercase : Any = pipe_prior( _lowerCAmelCase , generator=_lowerCAmelCase , num_inference_steps=5 , negative_prompt='' , ).to_tuple() _lowercase : Any = pipeline( _lowerCAmelCase , image=_lowerCAmelCase , image_embeds=_lowerCAmelCase , negative_image_embeds=_lowerCAmelCase , generator=_lowerCAmelCase , num_inference_steps=1_0_0 , height=7_6_8 , width=7_6_8 , strength=0.2 , output_type='np' , ) _lowercase : Union[str, Any] = output.images[0] assert image.shape == (7_6_8, 7_6_8, 3) assert_mean_pixel_difference(_lowerCAmelCase , _lowerCAmelCase )
66
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) UpperCamelCase = { "configuration_convnext": ["CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP", "ConvNextConfig", "ConvNextOnnxConfig"] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["ConvNextFeatureExtractor"] UpperCamelCase = ["ConvNextImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST", "ConvNextForImageClassification", "ConvNextModel", "ConvNextPreTrainedModel", "ConvNextBackbone", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "TFConvNextForImageClassification", "TFConvNextModel", "TFConvNextPreTrainedModel", ] if TYPE_CHECKING: from .configuration_convnext import CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvNextConfig, ConvNextOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_convnext import ConvNextFeatureExtractor from .image_processing_convnext import ConvNextImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convnext import ( CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvNextBackbone, ConvNextForImageClassification, ConvNextModel, ConvNextPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convnext import TFConvNextForImageClassification, TFConvNextModel, TFConvNextPreTrainedModel else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure)
66
1
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_bart import BartTokenizer UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} # See all BART models at https://huggingface.co/models?filter=bart UpperCamelCase = { "vocab_file": { "facebook/bart-base": "https://huggingface.co/facebook/bart-base/resolve/main/vocab.json", "facebook/bart-large": "https://huggingface.co/facebook/bart-large/resolve/main/vocab.json", "facebook/bart-large-mnli": "https://huggingface.co/facebook/bart-large-mnli/resolve/main/vocab.json", "facebook/bart-large-cnn": "https://huggingface.co/facebook/bart-large-cnn/resolve/main/vocab.json", "facebook/bart-large-xsum": "https://huggingface.co/facebook/bart-large-xsum/resolve/main/vocab.json", "yjernite/bart_eli5": "https://huggingface.co/yjernite/bart_eli5/resolve/main/vocab.json", }, "merges_file": { "facebook/bart-base": "https://huggingface.co/facebook/bart-base/resolve/main/merges.txt", "facebook/bart-large": "https://huggingface.co/facebook/bart-large/resolve/main/merges.txt", "facebook/bart-large-mnli": "https://huggingface.co/facebook/bart-large-mnli/resolve/main/merges.txt", "facebook/bart-large-cnn": "https://huggingface.co/facebook/bart-large-cnn/resolve/main/merges.txt", "facebook/bart-large-xsum": "https://huggingface.co/facebook/bart-large-xsum/resolve/main/merges.txt", "yjernite/bart_eli5": "https://huggingface.co/yjernite/bart_eli5/resolve/main/merges.txt", }, "tokenizer_file": { "facebook/bart-base": "https://huggingface.co/facebook/bart-base/resolve/main/tokenizer.json", "facebook/bart-large": "https://huggingface.co/facebook/bart-large/resolve/main/tokenizer.json", "facebook/bart-large-mnli": "https://huggingface.co/facebook/bart-large-mnli/resolve/main/tokenizer.json", "facebook/bart-large-cnn": "https://huggingface.co/facebook/bart-large-cnn/resolve/main/tokenizer.json", "facebook/bart-large-xsum": "https://huggingface.co/facebook/bart-large-xsum/resolve/main/tokenizer.json", "yjernite/bart_eli5": "https://huggingface.co/yjernite/bart_eli5/resolve/main/tokenizer.json", }, } UpperCamelCase = { "facebook/bart-base": 1_024, "facebook/bart-large": 1_024, "facebook/bart-large-mnli": 1_024, "facebook/bart-large-cnn": 1_024, "facebook/bart-large-xsum": 1_024, "yjernite/bart_eli5": 1_024, } class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : List[Any] = VOCAB_FILES_NAMES _UpperCamelCase : str = PRETRAINED_VOCAB_FILES_MAP _UpperCamelCase : Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCamelCase : Any = ["input_ids", "attention_mask"] _UpperCamelCase : List[str] = BartTokenizer def __init__( self , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase="replace" , _lowerCAmelCase="<s>" , _lowerCAmelCase="</s>" , _lowerCAmelCase="</s>" , _lowerCAmelCase="<s>" , _lowerCAmelCase="<unk>" , _lowerCAmelCase="<pad>" , _lowerCAmelCase="<mask>" , _lowerCAmelCase=False , _lowerCAmelCase=True , **_lowerCAmelCase , ): super().__init__( _lowerCAmelCase , _lowerCAmelCase , tokenizer_file=_lowerCAmelCase , errors=_lowerCAmelCase , bos_token=_lowerCAmelCase , eos_token=_lowerCAmelCase , sep_token=_lowerCAmelCase , cls_token=_lowerCAmelCase , unk_token=_lowerCAmelCase , pad_token=_lowerCAmelCase , mask_token=_lowerCAmelCase , add_prefix_space=_lowerCAmelCase , trim_offsets=_lowerCAmelCase , **_lowerCAmelCase , ) _lowercase : Tuple = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('add_prefix_space' , _lowerCAmelCase ) != add_prefix_space: _lowercase : List[Any] = getattr(_lowerCAmelCase , pre_tok_state.pop('type' ) ) _lowercase : str = add_prefix_space _lowercase : str = pre_tok_class(**_lowerCAmelCase ) _lowercase : Optional[int] = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` _lowercase : Tuple = 'post_processor' _lowercase : List[Any] = getattr(self.backend_tokenizer , _lowerCAmelCase , _lowerCAmelCase ) if tokenizer_component_instance: _lowercase : List[str] = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: _lowercase : Optional[Any] = tuple(state['sep'] ) if "cls" in state: _lowercase : Dict = tuple(state['cls'] ) _lowercase : Optional[int] = False if state.get('add_prefix_space' , _lowerCAmelCase ) != add_prefix_space: _lowercase : Optional[Any] = add_prefix_space _lowercase : Optional[Any] = True if state.get('trim_offsets' , _lowerCAmelCase ) != trim_offsets: _lowercase : Dict = trim_offsets _lowercase : Optional[int] = True if changes_to_apply: _lowercase : Any = getattr(_lowerCAmelCase , state.pop('type' ) ) _lowercase : Dict = component_class(**_lowerCAmelCase ) setattr(self.backend_tokenizer , _lowerCAmelCase , _lowerCAmelCase ) @property def __a ( self ): if self._mask_token is None: if self.verbose: logger.error('Using mask_token, but it is not set yet.' ) return None return str(self._mask_token ) @mask_token.setter def __a ( self , _lowerCAmelCase ): _lowercase : str = AddedToken(_lowerCAmelCase , lstrip=_lowerCAmelCase , rstrip=_lowerCAmelCase ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ) else value _lowercase : Dict = value def __a ( self , *_lowerCAmelCase , **_lowerCAmelCase ): _lowercase : Tuple = kwargs.get('is_split_into_words' , _lowerCAmelCase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True """ 'to use it with pretokenized inputs.' ) return super()._batch_encode_plus(*_lowerCAmelCase , **_lowerCAmelCase ) def __a ( self , *_lowerCAmelCase , **_lowerCAmelCase ): _lowercase : Optional[Any] = kwargs.get('is_split_into_words' , _lowerCAmelCase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True """ 'to use it with pretokenized inputs.' ) return super()._encode_plus(*_lowerCAmelCase , **_lowerCAmelCase ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase = None ): _lowercase : Tuple = self._tokenizer.model.save(_lowerCAmelCase , name=_lowerCAmelCase ) return tuple(_lowerCAmelCase ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase=None ): _lowercase : Union[str, Any] = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def __a ( self , _lowerCAmelCase , _lowerCAmelCase = None ): _lowercase : List[Any] = [self.sep_token_id] _lowercase : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
66
def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> bool: if p < 2: raise ValueError('p should not be less than 2!' ) elif p == 2: return True _lowercase : Optional[Any] = 4 _lowercase : Tuple = (1 << p) - 1 for _ in range(p - 2 ): _lowercase : Union[str, Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
66
1
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class lowerCAmelCase_ ( unittest.TestCase ): _UpperCamelCase : List[Any] = ViTImageProcessor if is_vision_available() else None @property def __a ( self ): return self.image_processor_tester.prepare_image_processor_dict() def __a ( self ): _lowercase : List[Any] = (3, 3_2, 1_2_8) _lowercase : List[Any] = tempfile.mkdtemp() # fmt: off _lowercase : str = ['[GO]', '[s]', '0', '1', '2', '3', '4', '5', '6', '7', '8', '9', 'a', 'b', 'c', 'd', 'e', 'f', 'g', 'h', 'i', 'j', 'k', 'l', 'm', 'n', 'o', 'p', 'q', 'r', 's', 't', 'u', 'v', 'w', 'x', 'y', 'z'] # fmt: on _lowercase : str = dict(zip(_lowerCAmelCase , range(len(_lowerCAmelCase ) ) ) ) _lowercase : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(_lowerCAmelCase ) + '\n' ) _lowercase : Dict = { 'do_normalize': False, 'do_resize': True, 'image_processor_type': 'ViTImageProcessor', 'resample': 3, 'size': {'height': 3_2, 'width': 1_2_8}, } _lowercase : int = os.path.join(self.tmpdirname , _lowerCAmelCase ) with open(self.image_processor_file , 'w' , encoding='utf-8' ) as fp: json.dump(_lowerCAmelCase , _lowerCAmelCase ) def __a ( self , **_lowerCAmelCase ): return MgpstrTokenizer.from_pretrained(self.tmpdirname , **_lowerCAmelCase ) def __a ( self , **_lowerCAmelCase ): return ViTImageProcessor.from_pretrained(self.tmpdirname , **_lowerCAmelCase ) def __a ( self ): shutil.rmtree(self.tmpdirname ) def __a ( self ): _lowercase : List[str] = np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta ) _lowercase : int = Image.fromarray(np.moveaxis(_lowerCAmelCase , 0 , -1 ) ) return image_input def __a ( self ): _lowercase : List[Any] = self.get_tokenizer() _lowercase : List[Any] = self.get_image_processor() _lowercase : str = MgpstrProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) processor.save_pretrained(self.tmpdirname ) _lowercase : Any = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=_lowerCAmelCase ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , _lowerCAmelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , _lowerCAmelCase ) def __a ( self ): _lowercase : Dict = self.get_tokenizer() _lowercase : int = self.get_image_processor() _lowercase : List[Any] = MgpstrProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) processor.save_pretrained(self.tmpdirname ) _lowercase : str = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) _lowercase : Optional[int] = self.get_image_processor(do_normalize=_lowerCAmelCase , padding_value=1.0 ) _lowercase : Any = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=_lowerCAmelCase , padding_value=1.0 ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , _lowerCAmelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , _lowerCAmelCase ) def __a ( self ): _lowercase : Optional[Any] = self.get_image_processor() _lowercase : Optional[int] = self.get_tokenizer() _lowercase : Optional[Any] = MgpstrProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) _lowercase : Tuple = self.prepare_image_inputs() _lowercase : str = image_processor(_lowerCAmelCase , return_tensors='np' ) _lowercase : Union[str, Any] = processor(images=_lowerCAmelCase , return_tensors='np' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1E-2 ) def __a ( self ): _lowercase : List[Any] = self.get_image_processor() _lowercase : Dict = self.get_tokenizer() _lowercase : str = MgpstrProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) _lowercase : Dict = 'test' _lowercase : Optional[int] = processor(text=_lowerCAmelCase ) _lowercase : Any = tokenizer(_lowerCAmelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def __a ( self ): _lowercase : int = self.get_image_processor() _lowercase : Tuple = self.get_tokenizer() _lowercase : int = MgpstrProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) _lowercase : Optional[Any] = 'test' _lowercase : str = self.prepare_image_inputs() _lowercase : Optional[Any] = processor(text=_lowerCAmelCase , images=_lowerCAmelCase ) self.assertListEqual(list(inputs.keys() ) , ['pixel_values', 'labels'] ) # test if it raises when no input is passed with pytest.raises(_lowerCAmelCase ): processor() def __a ( self ): _lowercase : Tuple = self.get_image_processor() _lowercase : Dict = self.get_tokenizer() _lowercase : Optional[Any] = MgpstrProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) _lowercase : Any = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] _lowercase : Optional[int] = processor.char_decode(_lowerCAmelCase ) _lowercase : List[Any] = tokenizer.batch_decode(_lowerCAmelCase ) _lowercase : Union[str, Any] = [seq.replace(' ' , '' ) for seq in decoded_tok] self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) def __a ( self ): _lowercase : Any = self.get_image_processor() _lowercase : List[str] = self.get_tokenizer() _lowercase : Union[str, Any] = MgpstrProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) _lowercase : Union[str, Any] = None _lowercase : Tuple = self.prepare_image_inputs() _lowercase : str = processor(text=_lowerCAmelCase , images=_lowerCAmelCase ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names ) def __a ( self ): _lowercase : Any = self.get_image_processor() _lowercase : int = self.get_tokenizer() _lowercase : Any = MgpstrProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) _lowercase : Union[str, Any] = torch.randn(1 , 2_7 , 3_8 ) _lowercase : Tuple = torch.randn(1 , 2_7 , 5_0_2_5_7 ) _lowercase : List[str] = torch.randn(1 , 2_7 , 3_0_5_2_2 ) _lowercase : Union[str, Any] = processor.batch_decode([char_input, bpe_input, wp_input] ) self.assertListEqual(list(results.keys() ) , ['generated_text', 'scores', 'char_preds', 'bpe_preds', 'wp_preds'] )
66
import random def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = False ) -> dict: _lowercase : dict = {i: [] for i in range(SCREAMING_SNAKE_CASE )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(SCREAMING_SNAKE_CASE ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(SCREAMING_SNAKE_CASE ): for j in range(i + 1 , SCREAMING_SNAKE_CASE ): if random.random() < probability: graph[i].append(SCREAMING_SNAKE_CASE ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(SCREAMING_SNAKE_CASE ) return graph def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> dict: return { i: [j for j in range(SCREAMING_SNAKE_CASE ) if i != j] for i in range(SCREAMING_SNAKE_CASE ) } if __name__ == "__main__": import doctest doctest.testmod()
66
1
import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : int = CodeGenTokenizer _UpperCamelCase : Any = CodeGenTokenizerFast _UpperCamelCase : Union[str, Any] = True _UpperCamelCase : Dict = {"add_prefix_space": True} _UpperCamelCase : Union[str, Any] = False def __a ( self ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt _lowercase : Optional[Any] = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', '<|endoftext|>', ] _lowercase : Any = dict(zip(_lowerCAmelCase , range(len(_lowerCAmelCase ) ) ) ) _lowercase : Dict = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] _lowercase : Any = {'unk_token': '<unk>'} _lowercase : int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) _lowercase : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(_lowerCAmelCase ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(_lowerCAmelCase ) ) def __a ( self , **_lowerCAmelCase ): kwargs.update(self.special_tokens_map ) return CodeGenTokenizer.from_pretrained(self.tmpdirname , **_lowerCAmelCase ) def __a ( self , **_lowerCAmelCase ): kwargs.update(self.special_tokens_map ) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname , **_lowerCAmelCase ) def __a ( self , _lowerCAmelCase ): _lowercase : Dict = 'lower newer' _lowercase : int = 'lower newer' return input_text, output_text def __a ( self ): _lowercase : Dict = CodeGenTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) _lowercase : int = 'lower newer' _lowercase : List[Any] = ['\u0120low', 'er', '\u0120', 'n', 'e', 'w', 'er'] _lowercase : int = tokenizer.tokenize(_lowerCAmelCase , add_prefix_space=_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) _lowercase : List[Any] = tokens + [tokenizer.unk_token] _lowercase : str = [1_4, 1_5, 1_0, 9, 3, 2, 1_5, 1_9] self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , _lowerCAmelCase ) def __a ( self ): if not self.test_rust_tokenizer: return _lowercase : Optional[int] = self.get_tokenizer() _lowercase : Dict = self.get_rust_tokenizer(add_prefix_space=_lowerCAmelCase ) _lowercase : Any = 'lower newer' # Testing tokenization _lowercase : List[str] = tokenizer.tokenize(_lowerCAmelCase , add_prefix_space=_lowerCAmelCase ) _lowercase : str = rust_tokenizer.tokenize(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) # Testing conversion to ids without special tokens _lowercase : Dict = tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase , add_prefix_space=_lowerCAmelCase ) _lowercase : Optional[int] = rust_tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) # Testing conversion to ids with special tokens _lowercase : Tuple = self.get_rust_tokenizer(add_prefix_space=_lowerCAmelCase ) _lowercase : List[Any] = tokenizer.encode(_lowerCAmelCase , add_prefix_space=_lowerCAmelCase ) _lowercase : Dict = rust_tokenizer.encode(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) # Testing the unknown token _lowercase : str = tokens + [rust_tokenizer.unk_token] _lowercase : str = [1_4, 1_5, 1_0, 9, 3, 2, 1_5, 1_9] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , _lowerCAmelCase ) def __a ( self , *_lowerCAmelCase , **_lowerCAmelCase ): # It's very difficult to mix/test pretokenization with byte-level # And get both CodeGen and Roberta to work at the same time (mostly an issue of adding a space before the string) pass def __a ( self , _lowerCAmelCase=1_5 ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _lowercase : str = self.rust_tokenizer_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) # Simple input _lowercase : Union[str, Any] = 'This is a simple input' _lowercase : Optional[int] = ['This is a simple input 1', 'This is a simple input 2'] _lowercase : Any = ('This is a simple input', 'This is a pair') _lowercase : Tuple = [ ('This is a simple input 1', 'This is a simple input 2'), ('This is a simple pair 1', 'This is a simple pair 2'), ] # Simple input tests self.assertRaises(_lowerCAmelCase , tokenizer_r.encode , _lowerCAmelCase , max_length=_lowerCAmelCase , padding='max_length' ) # Simple input self.assertRaises(_lowerCAmelCase , tokenizer_r.encode_plus , _lowerCAmelCase , max_length=_lowerCAmelCase , padding='max_length' ) # Simple input self.assertRaises( _lowerCAmelCase , tokenizer_r.batch_encode_plus , _lowerCAmelCase , max_length=_lowerCAmelCase , padding='max_length' , ) # Pair input self.assertRaises(_lowerCAmelCase , tokenizer_r.encode , _lowerCAmelCase , max_length=_lowerCAmelCase , padding='max_length' ) # Pair input self.assertRaises(_lowerCAmelCase , tokenizer_r.encode_plus , _lowerCAmelCase , max_length=_lowerCAmelCase , padding='max_length' ) # Pair input self.assertRaises( _lowerCAmelCase , tokenizer_r.batch_encode_plus , _lowerCAmelCase , max_length=_lowerCAmelCase , padding='max_length' , ) def __a ( self ): _lowercase : str = CodeGenTokenizer.from_pretrained(self.tmpdirname , pad_token='<pad>' ) # Simple input _lowercase : Union[str, Any] = 'This is a simple input' _lowercase : Any = ['This is a simple input looooooooong', 'This is a simple input'] _lowercase : Optional[int] = ('This is a simple input', 'This is a pair') _lowercase : Any = [ ('This is a simple input loooooong', 'This is a simple input'), ('This is a simple pair loooooong', 'This is a simple pair'), ] _lowercase : Union[str, Any] = tokenizer.pad_token_id _lowercase : str = tokenizer(_lowerCAmelCase , padding='max_length' , max_length=3_0 , return_tensors='np' ) _lowercase : str = tokenizer(_lowerCAmelCase , padding=_lowerCAmelCase , truncate=_lowerCAmelCase , return_tensors='np' ) _lowercase : List[str] = tokenizer(*_lowerCAmelCase , padding='max_length' , max_length=6_0 , return_tensors='np' ) _lowercase : int = tokenizer(_lowerCAmelCase , padding=_lowerCAmelCase , truncate=_lowerCAmelCase , return_tensors='np' ) # s # test single string max_length padding self.assertEqual(out_s['input_ids'].shape[-1] , 3_0 ) self.assertTrue(pad_token_id in out_s['input_ids'] ) self.assertTrue(0 in out_s['attention_mask'] ) # s2 # test automatic padding self.assertEqual(out_sa['input_ids'].shape[-1] , 3_3 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa['input_ids'][0] ) self.assertFalse(0 in out_sa['attention_mask'][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa['input_ids'][1] ) self.assertTrue(0 in out_sa['attention_mask'][1] ) # p # test single pair max_length padding self.assertEqual(out_p['input_ids'].shape[-1] , 6_0 ) self.assertTrue(pad_token_id in out_p['input_ids'] ) self.assertTrue(0 in out_p['attention_mask'] ) # p2 # test automatic padding pair self.assertEqual(out_pa['input_ids'].shape[-1] , 5_2 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa['input_ids'][0] ) self.assertFalse(0 in out_pa['attention_mask'][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa['input_ids'][1] ) self.assertTrue(0 in out_pa['attention_mask'][1] ) def __a ( self ): _lowercase : str = '$$$' _lowercase : int = CodeGenTokenizer.from_pretrained(self.tmpdirname , bos_token=_lowerCAmelCase , add_bos_token=_lowerCAmelCase ) _lowercase : Any = 'This is a simple input' _lowercase : List[Any] = ['This is a simple input 1', 'This is a simple input 2'] _lowercase : List[str] = tokenizer.bos_token_id _lowercase : int = tokenizer(_lowerCAmelCase ) _lowercase : str = tokenizer(_lowerCAmelCase ) self.assertEqual(out_s.input_ids[0] , _lowerCAmelCase ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) _lowercase : Dict = tokenizer.decode(out_s.input_ids ) _lowercase : List[Any] = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , _lowerCAmelCase ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) @slow def __a ( self ): _lowercase : List[str] = CodeGenTokenizer.from_pretrained('Salesforce/codegen-350M-mono' ) _lowercase : Union[str, Any] = '\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#' _lowercase : str = '\nif len_a > len_b: result = a\nelse: result = b' _lowercase : Optional[Any] = tokenizer.encode(_lowerCAmelCase ) _lowercase : Optional[Any] = ['^#', re.escape('<|endoftext|>' ), '^\'\'\'', '^"""', '\n\n\n'] _lowercase : Tuple = tokenizer.decode(_lowerCAmelCase , truncate_before_pattern=_lowerCAmelCase ) self.assertEqual(_lowerCAmelCase , _lowerCAmelCase ) def __a ( self ): pass
66
from __future__ import annotations UpperCamelCase = tuple[int, int, int] UpperCamelCase = tuple[str, str, str] # used alphabet -------------------------- # from string.ascii_uppercase UpperCamelCase = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" # -------------------------- default selection -------------------------- # rotors -------------------------- UpperCamelCase = "EGZWVONAHDCLFQMSIPJBYUKXTR" UpperCamelCase = "FOBHMDKEXQNRAULPGSJVTYICZW" UpperCamelCase = "ZJXESIUQLHAVRMDOYGTNFWPBKC" # reflector -------------------------- UpperCamelCase = { "A": "N", "N": "A", "B": "O", "O": "B", "C": "P", "P": "C", "D": "Q", "Q": "D", "E": "R", "R": "E", "F": "S", "S": "F", "G": "T", "T": "G", "H": "U", "U": "H", "I": "V", "V": "I", "J": "W", "W": "J", "K": "X", "X": "K", "L": "Y", "Y": "L", "M": "Z", "Z": "M", } # -------------------------- extra rotors -------------------------- UpperCamelCase = "RMDJXFUWGISLHVTCQNKYPBEZOA" UpperCamelCase = "SGLCPQWZHKXAREONTFBVIYJUDM" UpperCamelCase = "HVSICLTYKQUBXDWAJZOMFGPREN" UpperCamelCase = "RZWQHFMVDBKICJLNTUXAGYPSOE" UpperCamelCase = "LFKIJODBEGAMQPXVUHYSTCZRWN" UpperCamelCase = "KOAEGVDHXPQZMLFTYWJNBRCIUS" def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> tuple[RotorPositionT, RotorSelectionT, dict[str, str]]: # Checks if there are 3 unique rotors if (unique_rotsel := len(set(SCREAMING_SNAKE_CASE ) )) < 3: _lowercase : Optional[int] = F"""Please use 3 unique rotors (not {unique_rotsel})""" raise Exception(SCREAMING_SNAKE_CASE ) # Checks if rotor positions are valid _lowercase , _lowercase , _lowercase : int = rotpos if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : Dict = F"""First rotor position is not within range of 1..26 ({rotorposa}""" raise ValueError(SCREAMING_SNAKE_CASE ) if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : int = F"""Second rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(SCREAMING_SNAKE_CASE ) if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : str = F"""Third rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(SCREAMING_SNAKE_CASE ) # Validates string and returns dict _lowercase : Tuple = _plugboard(SCREAMING_SNAKE_CASE ) return rotpos, rotsel, pbdict def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> dict[str, str]: # tests the input string if it # a) is type string # b) has even length (so pairs can be made) if not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): _lowercase : Optional[int] = F"""Plugboard setting isn't type string ({type(SCREAMING_SNAKE_CASE )})""" raise TypeError(SCREAMING_SNAKE_CASE ) elif len(SCREAMING_SNAKE_CASE ) % 2 != 0: _lowercase : Optional[int] = F"""Odd number of symbols ({len(SCREAMING_SNAKE_CASE )})""" raise Exception(SCREAMING_SNAKE_CASE ) elif pbstring == "": return {} pbstring.replace(' ' , '' ) # Checks if all characters are unique _lowercase : Dict = set() for i in pbstring: if i not in abc: _lowercase : str = F"""'{i}' not in list of symbols""" raise Exception(SCREAMING_SNAKE_CASE ) elif i in tmppbl: _lowercase : int = F"""Duplicate symbol ({i})""" raise Exception(SCREAMING_SNAKE_CASE ) else: tmppbl.add(SCREAMING_SNAKE_CASE ) del tmppbl # Created the dictionary _lowercase : Optional[Any] = {} for j in range(0 , len(SCREAMING_SNAKE_CASE ) - 1 , 2 ): _lowercase : Dict = pbstring[j + 1] _lowercase : Union[str, Any] = pbstring[j] return pb def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = (rotora, rotora, rotora) , SCREAMING_SNAKE_CASE = "" , ) -> str: _lowercase : List[str] = text.upper() _lowercase , _lowercase , _lowercase : List[str] = _validator( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , plugb.upper() ) _lowercase , _lowercase , _lowercase : Optional[int] = rotor_position _lowercase , _lowercase , _lowercase : Union[str, Any] = rotor_selection rotorposa -= 1 rotorposa -= 1 rotorposa -= 1 _lowercase : Optional[int] = [] # encryption/decryption process -------------------------- for symbol in text: if symbol in abc: # 1st plugboard -------------------------- if symbol in plugboard: _lowercase : Dict = plugboard[symbol] # rotor ra -------------------------- _lowercase : Optional[Any] = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : Union[str, Any] = rotora[index % len(SCREAMING_SNAKE_CASE )] # rotor rb -------------------------- _lowercase : Tuple = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : str = rotora[index % len(SCREAMING_SNAKE_CASE )] # rotor rc -------------------------- _lowercase : List[Any] = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : List[str] = rotora[index % len(SCREAMING_SNAKE_CASE )] # reflector -------------------------- # this is the reason you don't need another machine to decipher _lowercase : List[str] = reflector[symbol] # 2nd rotors _lowercase : List[str] = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] _lowercase : Tuple = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] _lowercase : Dict = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] # 2nd plugboard if symbol in plugboard: _lowercase : int = plugboard[symbol] # moves/resets rotor positions rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : Any = 0 rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : int = 0 rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : Any = 0 # else: # pass # Error could be also raised # raise ValueError( # 'Invalid symbol('+repr(symbol)+')') result.append(SCREAMING_SNAKE_CASE ) return "".join(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": UpperCamelCase = "This is my Python script that emulates the Enigma machine from WWII." UpperCamelCase = (1, 1, 1) UpperCamelCase = "pictures" UpperCamelCase = (rotora, rotora, rotora) UpperCamelCase = enigma(message, rotor_pos, rotor_sel, pb) print("Encrypted message:", en) print("Decrypted message:", enigma(en, rotor_pos, rotor_sel, pb))
66
1
import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, PLBartTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin UpperCamelCase = get_tests_dir("fixtures/test_sentencepiece.model") if is_torch_available(): from transformers.models.plbart.modeling_plbart import shift_tokens_right UpperCamelCase = 50_003 UpperCamelCase = 50_002 @require_sentencepiece @require_tokenizers class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : List[Any] = PLBartTokenizer _UpperCamelCase : Tuple = None _UpperCamelCase : int = False def __a ( self ): super().setUp() # We have a SentencePiece fixture for testing _lowercase : List[Any] = PLBartTokenizer(_lowerCAmelCase , language_codes='base' , keep_accents=_lowerCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def __a ( self ): _lowercase : Tuple = PLBartTokenizer(_lowerCAmelCase , language_codes='base' , keep_accents=_lowerCAmelCase ) _lowercase : Dict = tokenizer.tokenize('This is a test' ) self.assertListEqual(_lowerCAmelCase , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] , ) _lowercase : Any = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) _lowercase : Optional[Any] = tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) self.assertListEqual( _lowerCAmelCase , [ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 2, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 2, 4] ] , ) _lowercase : Tuple = tokenizer.convert_ids_to_tokens(_lowerCAmelCase ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) _lowercase : str = tokenizer.vocab_size _lowercase : List[Any] = [tokenizer.convert_ids_to_tokens(_lowerCAmelCase ) for x in range(end - 4 , _lowerCAmelCase )] self.assertListEqual(_lowerCAmelCase , ['__java__', '__python__', '__en_XX__', '<mask>'] ) _lowercase : Optional[Any] = 'java.lang.Exception, python.lang.Exception, javascript, php, ruby, go' _lowercase : str = tokenizer(_lowerCAmelCase ).input_ids self.assertEqual( tokenizer.decode(_lowerCAmelCase , skip_special_tokens=_lowerCAmelCase , clean_up_tokenization_spaces=_lowerCAmelCase ) , _lowerCAmelCase , ) def __a ( self ): _lowercase : Union[str, Any] = PLBartTokenizer(_lowerCAmelCase , language_codes='multi' , keep_accents=_lowerCAmelCase ) _lowercase : Dict = tokenizer.tokenize('This is a test' ) self.assertListEqual(_lowerCAmelCase , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] , ) _lowercase : str = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) _lowercase : Dict = tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) self.assertListEqual( _lowerCAmelCase , [ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 2, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 2, 4] ] , ) _lowercase : Any = tokenizer.convert_ids_to_tokens(_lowerCAmelCase ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) _lowercase : int = tokenizer.vocab_size _lowercase : Union[str, Any] = [tokenizer.convert_ids_to_tokens(_lowerCAmelCase ) for x in range(end - 7 , _lowerCAmelCase )] self.assertListEqual( _lowerCAmelCase , ['__java__', '__python__', '__en_XX__', '__javascript__', '__php__', '__ruby__', '__go__'] ) _lowercase : Any = 'java.lang.Exception, python.lang.Exception, javascript, php, ruby, go' _lowercase : List[Any] = tokenizer(_lowerCAmelCase ).input_ids self.assertEqual( tokenizer.decode(_lowerCAmelCase , skip_special_tokens=_lowerCAmelCase , clean_up_tokenization_spaces=_lowerCAmelCase ) , _lowerCAmelCase , ) @require_torch @require_sentencepiece @require_tokenizers class lowerCAmelCase_ ( unittest.TestCase ): _UpperCamelCase : Optional[Any] = "uclanlp/plbart-python-en_XX" _UpperCamelCase : Dict = [ "def maximum(a,b,c):NEW_LINE_INDENTreturn max([a,b,c])", "def sum(a,b,c):NEW_LINE_INDENTreturn sum([a,b,c])", ] _UpperCamelCase : List[str] = [ "Returns the maximum value of a b c.", "Sums the values of a b c.", ] _UpperCamelCase : Optional[Any] = [ 134, 5452, 33460, 33441, 33463, 33465, 33463, 33449, 988, 20, 33456, 19, 33456, 771, 39, 4258, 889, 3318, 33441, 33463, 33465, 33463, 33449, 2471, 2, PYTHON_CODE, ] @classmethod def __a ( cls ): _lowercase : PLBartTokenizer = PLBartTokenizer.from_pretrained( cls.checkpoint_name , language_codes='base' , src_lang='python' , tgt_lang='en_XX' ) _lowercase : Optional[Any] = 1 return cls def __a ( self ): self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['__java__'] , 5_0_0_0_1 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['__python__'] , 5_0_0_0_2 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['__en_XX__'] , 5_0_0_0_3 ) def __a ( self ): _lowercase : Any = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , _lowerCAmelCase ) def __a ( self ): self.assertIn(_lowerCAmelCase , self.tokenizer.all_special_ids ) _lowercase : Any = [EN_CODE, 9_0_3_7, 3_3_4_4_2, 5_7, 7_5_2, 1_5_3, 1_4, 5_6, 1_8, 9, 2] _lowercase : Optional[int] = self.tokenizer.decode(_lowerCAmelCase , skip_special_tokens=_lowerCAmelCase ) _lowercase : str = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=_lowerCAmelCase ) self.assertEqual(_lowerCAmelCase , _lowerCAmelCase ) self.assertNotIn(self.tokenizer.eos_token , _lowerCAmelCase ) def __a ( self ): _lowercase : Optional[Any] = ['def sum(a,b,c):NEW_LINE_INDENTreturn sum([a,b,c])' * 2_0] self.assertIsInstance(src_text[0] , _lowerCAmelCase ) _lowercase : List[str] = 1_0 _lowercase : List[str] = self.tokenizer(_lowerCAmelCase , max_length=_lowerCAmelCase , truncation=_lowerCAmelCase ).input_ids[0] self.assertEqual(ids[-2] , 2 ) self.assertEqual(ids[-1] , _lowerCAmelCase ) self.assertEqual(len(_lowerCAmelCase ) , _lowerCAmelCase ) def __a ( self ): self.assertListEqual(self.tokenizer.convert_tokens_to_ids(['<mask>', '__java__'] ) , [5_0_0_0_4, 5_0_0_0_1] ) def __a ( self ): _lowercase : Tuple = tempfile.mkdtemp() _lowercase : Tuple = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(_lowerCAmelCase ) _lowercase : Any = PLBartTokenizer.from_pretrained(_lowerCAmelCase ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , _lowerCAmelCase ) @require_torch def __a ( self ): _lowercase : List[Any] = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=_lowerCAmelCase , return_tensors='pt' ) _lowercase : Optional[int] = shift_tokens_right(batch['labels'] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 self.assertEqual(batch.input_ids[1][-2:].tolist() , [2, PYTHON_CODE] ) self.assertEqual(batch.decoder_input_ids[1][0] , _lowerCAmelCase ) self.assertEqual(batch.decoder_input_ids[1][-1] , 2 ) self.assertEqual(batch.labels[1][-2:].tolist() , [2, EN_CODE] ) @require_torch def __a ( self ): _lowercase : int = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=_lowerCAmelCase , truncation=_lowerCAmelCase , max_length=len(self.expected_src_tokens ) , return_tensors='pt' , ) _lowercase : List[str] = shift_tokens_right(batch['labels'] , self.tokenizer.pad_token_id ) self.assertIsInstance(_lowerCAmelCase , _lowerCAmelCase ) self.assertEqual((2, 2_6) , batch.input_ids.shape ) self.assertEqual((2, 2_6) , batch.attention_mask.shape ) _lowercase : Tuple = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , _lowerCAmelCase ) self.assertEqual(2 , batch.decoder_input_ids[0, -1] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, PYTHON_CODE] ) def __a ( self ): _lowercase : List[str] = self.tokenizer(self.src_text , padding=_lowerCAmelCase , truncation=_lowerCAmelCase , max_length=3 , return_tensors='pt' ) _lowercase : int = self.tokenizer( text_target=self.tgt_text , padding=_lowerCAmelCase , truncation=_lowerCAmelCase , max_length=1_0 , return_tensors='pt' ) _lowercase : Any = targets['input_ids'] _lowercase : Optional[int] = shift_tokens_right(_lowerCAmelCase , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 1_0 ) @require_torch def __a ( self ): _lowercase : Tuple = self.tokenizer._build_translation_inputs( 'A test' , return_tensors='pt' , src_lang='en_XX' , tgt_lang='java' ) self.assertEqual( nested_simplify(_lowerCAmelCase ) , { # A, test, EOS, en_XX 'input_ids': [[1_5_0, 2_4_2, 2, 5_0_0_0_3]], 'attention_mask': [[1, 1, 1, 1]], # java 'forced_bos_token_id': 5_0_0_0_1, } , )
66
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = {"configuration_glpn": ["GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP", "GLPNConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["GLPNFeatureExtractor"] UpperCamelCase = ["GLPNImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "GLPN_PRETRAINED_MODEL_ARCHIVE_LIST", "GLPNForDepthEstimation", "GLPNLayer", "GLPNModel", "GLPNPreTrainedModel", ] if TYPE_CHECKING: from .configuration_glpn import GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP, GLPNConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_glpn import GLPNFeatureExtractor from .image_processing_glpn import GLPNImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_glpn import ( GLPN_PRETRAINED_MODEL_ARCHIVE_LIST, GLPNForDepthEstimation, GLPNLayer, GLPNModel, GLPNPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) UpperCamelCase = { "configuration_efficientformer": [ "EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "EfficientFormerConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["EfficientFormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "EfficientFormerForImageClassification", "EfficientFormerForImageClassificationWithTeacher", "EfficientFormerModel", "EfficientFormerPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "TFEfficientFormerForImageClassification", "TFEfficientFormerForImageClassificationWithTeacher", "TFEfficientFormerModel", "TFEfficientFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_efficientformer import EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientFormerConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientformer import EfficientFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientformer import ( EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientFormerForImageClassification, EfficientFormerForImageClassificationWithTeacher, EfficientFormerModel, EfficientFormerPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, TFEfficientFormerPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
import unittest from transformers import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device if is_torch_available(): import torch from transformers import AutoModelForImageClassification if is_vision_available(): from transformers import AutoImageProcessor @require_torch @require_vision class lowerCAmelCase_ ( unittest.TestCase ): @slow def __a ( self ): _lowercase : List[Any] = AutoImageProcessor.from_pretrained('microsoft/dit-base-finetuned-rvlcdip' ) _lowercase : List[Any] = AutoModelForImageClassification.from_pretrained('microsoft/dit-base-finetuned-rvlcdip' ) model.to(_lowerCAmelCase ) from datasets import load_dataset _lowercase : Union[str, Any] = load_dataset('nielsr/rvlcdip-demo' ) _lowercase : Any = dataset['train'][0]['image'].convert('RGB' ) _lowercase : List[str] = image_processor(_lowerCAmelCase , return_tensors='pt' ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): _lowercase : Dict = model(**_lowerCAmelCase ) _lowercase : Any = outputs.logits _lowercase : str = torch.Size((1, 1_6) ) self.assertEqual(logits.shape , _lowerCAmelCase ) _lowercase : Union[str, Any] = torch.tensor( [-0.41_58, -0.40_92, -0.43_47] , device=_lowerCAmelCase , dtype=torch.float , ) self.assertTrue(torch.allclose(logits[0, :3] , _lowerCAmelCase , atol=1E-4 ) )
66
1
import argparse import os import torch from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) UpperCamelCase = { "sample_size": 32, "in_channels": 3, "out_channels": 3, "layers_per_block": 2, "num_class_embeds": 1_000, "block_out_channels": [32, 64], "attention_head_dim": 8, "down_block_types": [ "ResnetDownsampleBlock2D", "AttnDownBlock2D", ], "up_block_types": [ "AttnUpBlock2D", "ResnetUpsampleBlock2D", ], "resnet_time_scale_shift": "scale_shift", "upsample_type": "resnet", "downsample_type": "resnet", } UpperCamelCase = { "sample_size": 64, "in_channels": 3, "out_channels": 3, "layers_per_block": 3, "num_class_embeds": 1_000, "block_out_channels": [192, 192 * 2, 192 * 3, 192 * 4], "attention_head_dim": 64, "down_block_types": [ "ResnetDownsampleBlock2D", "AttnDownBlock2D", "AttnDownBlock2D", "AttnDownBlock2D", ], "up_block_types": [ "AttnUpBlock2D", "AttnUpBlock2D", "AttnUpBlock2D", "ResnetUpsampleBlock2D", ], "resnet_time_scale_shift": "scale_shift", "upsample_type": "resnet", "downsample_type": "resnet", } UpperCamelCase = { "sample_size": 256, "in_channels": 3, "out_channels": 3, "layers_per_block": 2, "num_class_embeds": None, "block_out_channels": [256, 256, 256 * 2, 256 * 2, 256 * 4, 256 * 4], "attention_head_dim": 64, "down_block_types": [ "ResnetDownsampleBlock2D", "ResnetDownsampleBlock2D", "ResnetDownsampleBlock2D", "AttnDownBlock2D", "AttnDownBlock2D", "AttnDownBlock2D", ], "up_block_types": [ "AttnUpBlock2D", "AttnUpBlock2D", "AttnUpBlock2D", "ResnetUpsampleBlock2D", "ResnetUpsampleBlock2D", "ResnetUpsampleBlock2D", ], "resnet_time_scale_shift": "default", "upsample_type": "resnet", "downsample_type": "resnet", } UpperCamelCase = { "num_train_timesteps": 40, "sigma_min": 0.0_02, "sigma_max": 80.0, } UpperCamelCase = { "num_train_timesteps": 201, "sigma_min": 0.0_02, "sigma_max": 80.0, } UpperCamelCase = { "num_train_timesteps": 151, "sigma_min": 0.0_02, "sigma_max": 80.0, } def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> Dict: if isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise argparse.ArgumentTypeError('boolean value expected' ) def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=False ) -> Tuple: _lowercase : List[str] = checkpoint[F"""{old_prefix}.in_layers.0.weight"""] _lowercase : int = checkpoint[F"""{old_prefix}.in_layers.0.bias"""] _lowercase : Optional[int] = checkpoint[F"""{old_prefix}.in_layers.2.weight"""] _lowercase : Optional[Any] = checkpoint[F"""{old_prefix}.in_layers.2.bias"""] _lowercase : Optional[Any] = checkpoint[F"""{old_prefix}.emb_layers.1.weight"""] _lowercase : Optional[int] = checkpoint[F"""{old_prefix}.emb_layers.1.bias"""] _lowercase : Union[str, Any] = checkpoint[F"""{old_prefix}.out_layers.0.weight"""] _lowercase : List[str] = checkpoint[F"""{old_prefix}.out_layers.0.bias"""] _lowercase : Optional[int] = checkpoint[F"""{old_prefix}.out_layers.3.weight"""] _lowercase : Optional[Any] = checkpoint[F"""{old_prefix}.out_layers.3.bias"""] if has_skip: _lowercase : Union[str, Any] = checkpoint[F"""{old_prefix}.skip_connection.weight"""] _lowercase : int = checkpoint[F"""{old_prefix}.skip_connection.bias"""] return new_checkpoint def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=None ) -> List[str]: _lowercase , _lowercase , _lowercase : Union[str, Any] = checkpoint[F"""{old_prefix}.qkv.weight"""].chunk(3 , dim=0 ) _lowercase , _lowercase , _lowercase : Union[str, Any] = checkpoint[F"""{old_prefix}.qkv.bias"""].chunk(3 , dim=0 ) _lowercase : str = checkpoint[F"""{old_prefix}.norm.weight"""] _lowercase : Tuple = checkpoint[F"""{old_prefix}.norm.bias"""] _lowercase : int = weight_q.squeeze(-1 ).squeeze(-1 ) _lowercase : List[Any] = bias_q.squeeze(-1 ).squeeze(-1 ) _lowercase : int = weight_k.squeeze(-1 ).squeeze(-1 ) _lowercase : Any = bias_k.squeeze(-1 ).squeeze(-1 ) _lowercase : List[str] = weight_v.squeeze(-1 ).squeeze(-1 ) _lowercase : List[str] = bias_v.squeeze(-1 ).squeeze(-1 ) _lowercase : Tuple = ( checkpoint[F"""{old_prefix}.proj_out.weight"""].squeeze(-1 ).squeeze(-1 ) ) _lowercase : Dict = checkpoint[F"""{old_prefix}.proj_out.bias"""].squeeze(-1 ).squeeze(-1 ) return new_checkpoint def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: _lowercase : str = torch.load(SCREAMING_SNAKE_CASE , map_location='cpu' ) _lowercase : str = {} _lowercase : str = checkpoint['time_embed.0.weight'] _lowercase : Tuple = checkpoint['time_embed.0.bias'] _lowercase : Optional[Any] = checkpoint['time_embed.2.weight'] _lowercase : int = checkpoint['time_embed.2.bias'] if unet_config["num_class_embeds"] is not None: _lowercase : Dict = checkpoint['label_emb.weight'] _lowercase : Tuple = checkpoint['input_blocks.0.0.weight'] _lowercase : Optional[Any] = checkpoint['input_blocks.0.0.bias'] _lowercase : Union[str, Any] = unet_config['down_block_types'] _lowercase : Tuple = unet_config['layers_per_block'] _lowercase : Any = unet_config['attention_head_dim'] _lowercase : List[Any] = unet_config['block_out_channels'] _lowercase : Optional[Any] = 1 _lowercase : List[str] = channels_list[0] for i, layer_type in enumerate(SCREAMING_SNAKE_CASE ): _lowercase : Tuple = channels_list[i] _lowercase : Tuple = current_channels != prev_channels if layer_type == "ResnetDownsampleBlock2D": for j in range(SCREAMING_SNAKE_CASE ): _lowercase : str = F"""down_blocks.{i}.resnets.{j}""" _lowercase : Dict = F"""input_blocks.{current_layer}.0""" _lowercase : int = True if j == 0 and downsample_block_has_skip else False _lowercase : Optional[int] = convert_resnet(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , has_skip=SCREAMING_SNAKE_CASE ) current_layer += 1 elif layer_type == "AttnDownBlock2D": for j in range(SCREAMING_SNAKE_CASE ): _lowercase : Tuple = F"""down_blocks.{i}.resnets.{j}""" _lowercase : Union[str, Any] = F"""input_blocks.{current_layer}.0""" _lowercase : List[Any] = True if j == 0 and downsample_block_has_skip else False _lowercase : Union[str, Any] = convert_resnet(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , has_skip=SCREAMING_SNAKE_CASE ) _lowercase : Dict = F"""down_blocks.{i}.attentions.{j}""" _lowercase : int = F"""input_blocks.{current_layer}.1""" _lowercase : Any = convert_attention( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) current_layer += 1 if i != len(SCREAMING_SNAKE_CASE ) - 1: _lowercase : int = F"""down_blocks.{i}.downsamplers.0""" _lowercase : List[Any] = F"""input_blocks.{current_layer}.0""" _lowercase : Tuple = convert_resnet(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) current_layer += 1 _lowercase : List[Any] = current_channels # hardcoded the mid-block for now _lowercase : List[Any] = 'mid_block.resnets.0' _lowercase : Union[str, Any] = 'middle_block.0' _lowercase : Optional[int] = convert_resnet(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) _lowercase : Tuple = 'mid_block.attentions.0' _lowercase : Any = 'middle_block.1' _lowercase : str = convert_attention(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) _lowercase : Optional[Any] = 'mid_block.resnets.1' _lowercase : Optional[Any] = 'middle_block.2' _lowercase : List[Any] = convert_resnet(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) _lowercase : Optional[int] = 0 _lowercase : List[str] = unet_config['up_block_types'] for i, layer_type in enumerate(SCREAMING_SNAKE_CASE ): if layer_type == "ResnetUpsampleBlock2D": for j in range(layers_per_block + 1 ): _lowercase : Tuple = F"""up_blocks.{i}.resnets.{j}""" _lowercase : int = F"""output_blocks.{current_layer}.0""" _lowercase : str = convert_resnet(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , has_skip=SCREAMING_SNAKE_CASE ) current_layer += 1 if i != len(SCREAMING_SNAKE_CASE ) - 1: _lowercase : Dict = F"""up_blocks.{i}.upsamplers.0""" _lowercase : Optional[Any] = F"""output_blocks.{current_layer-1}.1""" _lowercase : Dict = convert_resnet(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) elif layer_type == "AttnUpBlock2D": for j in range(layers_per_block + 1 ): _lowercase : Tuple = F"""up_blocks.{i}.resnets.{j}""" _lowercase : List[str] = F"""output_blocks.{current_layer}.0""" _lowercase : Optional[Any] = convert_resnet(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , has_skip=SCREAMING_SNAKE_CASE ) _lowercase : str = F"""up_blocks.{i}.attentions.{j}""" _lowercase : Dict = F"""output_blocks.{current_layer}.1""" _lowercase : Optional[Any] = convert_attention( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) current_layer += 1 if i != len(SCREAMING_SNAKE_CASE ) - 1: _lowercase : int = F"""up_blocks.{i}.upsamplers.0""" _lowercase : str = F"""output_blocks.{current_layer-1}.2""" _lowercase : List[Any] = convert_resnet(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) _lowercase : List[Any] = checkpoint['out.0.weight'] _lowercase : Optional[Any] = checkpoint['out.0.bias'] _lowercase : Dict = checkpoint['out.2.weight'] _lowercase : Dict = checkpoint['out.2.bias'] return new_checkpoint if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() parser.add_argument("--unet_path", default=None, type=str, required=True, help="Path to the unet.pt to convert.") parser.add_argument( "--dump_path", default=None, type=str, required=True, help="Path to output the converted UNet model." ) parser.add_argument("--class_cond", default=True, type=str, help="Whether the model is class-conditional.") UpperCamelCase = parser.parse_args() UpperCamelCase = strabool(args.class_cond) UpperCamelCase = os.path.basename(args.unet_path) print(f'''Checkpoint: {ckpt_name}''') # Get U-Net config if "imagenet64" in ckpt_name: UpperCamelCase = IMAGENET_64_UNET_CONFIG elif "256" in ckpt_name and (("bedroom" in ckpt_name) or ("cat" in ckpt_name)): UpperCamelCase = LSUN_256_UNET_CONFIG elif "test" in ckpt_name: UpperCamelCase = TEST_UNET_CONFIG else: raise ValueError(f'''Checkpoint type {ckpt_name} is not currently supported.''') if not args.class_cond: UpperCamelCase = None UpperCamelCase = con_pt_to_diffuser(args.unet_path, unet_config) UpperCamelCase = UNetaDModel(**unet_config) image_unet.load_state_dict(converted_unet_ckpt) # Get scheduler config if "cd" in ckpt_name or "test" in ckpt_name: UpperCamelCase = CD_SCHEDULER_CONFIG elif "ct" in ckpt_name and "imagenet64" in ckpt_name: UpperCamelCase = CT_IMAGENET_64_SCHEDULER_CONFIG elif "ct" in ckpt_name and "256" in ckpt_name and (("bedroom" in ckpt_name) or ("cat" in ckpt_name)): UpperCamelCase = CT_LSUN_256_SCHEDULER_CONFIG else: raise ValueError(f'''Checkpoint type {ckpt_name} is not currently supported.''') UpperCamelCase = CMStochasticIterativeScheduler(**scheduler_config) UpperCamelCase = ConsistencyModelPipeline(unet=image_unet, scheduler=cm_scheduler) consistency_model.save_pretrained(args.dump_path)
66
from PIL import Image def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Image: def brightness(SCREAMING_SNAKE_CASE ) -> float: return 128 + level + (c - 128) if not -255.0 <= level <= 255.0: raise ValueError('level must be between -255.0 (black) and 255.0 (white)' ) return img.point(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": # Load image with Image.open("image_data/lena.jpg") as img: # Change brightness to 100 UpperCamelCase = change_brightness(img, 100) brigt_img.save("image_data/lena_brightness.png", format="png")
66
1
import copy import unittest from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_MULTIPLE_CHOICE_MAPPING, MODEL_FOR_QUESTION_ANSWERING_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, LayoutLMvaConfig, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, ) from transformers.models.layoutlmva.modeling_layoutlmva import LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class lowerCAmelCase_ : def __init__( self , _lowerCAmelCase , _lowerCAmelCase=2 , _lowerCAmelCase=3 , _lowerCAmelCase=4 , _lowerCAmelCase=2 , _lowerCAmelCase=7 , _lowerCAmelCase=True , _lowerCAmelCase=True , _lowerCAmelCase=True , _lowerCAmelCase=True , _lowerCAmelCase=9_9 , _lowerCAmelCase=3_6 , _lowerCAmelCase=3 , _lowerCAmelCase=4 , _lowerCAmelCase=3_7 , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.1 , _lowerCAmelCase=5_1_2 , _lowerCAmelCase=1_6 , _lowerCAmelCase=2 , _lowerCAmelCase=0.02 , _lowerCAmelCase=6 , _lowerCAmelCase=6 , _lowerCAmelCase=3 , _lowerCAmelCase=4 , _lowerCAmelCase=None , _lowerCAmelCase=1_0_0_0 , ): _lowercase : Optional[Any] = parent _lowercase : Optional[int] = batch_size _lowercase : List[str] = num_channels _lowercase : Tuple = image_size _lowercase : str = patch_size _lowercase : Union[str, Any] = text_seq_length _lowercase : List[str] = is_training _lowercase : Optional[int] = use_input_mask _lowercase : Dict = use_token_type_ids _lowercase : Any = use_labels _lowercase : Optional[int] = vocab_size _lowercase : int = hidden_size _lowercase : Dict = num_hidden_layers _lowercase : Optional[Any] = num_attention_heads _lowercase : List[Any] = intermediate_size _lowercase : Optional[int] = hidden_act _lowercase : Dict = hidden_dropout_prob _lowercase : List[str] = attention_probs_dropout_prob _lowercase : Union[str, Any] = max_position_embeddings _lowercase : Optional[Any] = type_vocab_size _lowercase : List[str] = type_sequence_label_size _lowercase : Union[str, Any] = initializer_range _lowercase : Optional[Any] = coordinate_size _lowercase : Tuple = shape_size _lowercase : Any = num_labels _lowercase : List[str] = num_choices _lowercase : Optional[Any] = scope _lowercase : str = range_bbox # LayoutLMv3's sequence length equals the number of text tokens + number of patches + 1 (we add 1 for the CLS token) _lowercase : Optional[Any] = text_seq_length _lowercase : Dict = (image_size // patch_size) ** 2 + 1 _lowercase : Union[str, Any] = self.text_seq_length + self.image_seq_length def __a ( self ): _lowercase : str = ids_tensor([self.batch_size, self.text_seq_length] , self.vocab_size ) _lowercase : Dict = ids_tensor([self.batch_size, self.text_seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: _lowercase : Optional[int] = bbox[i, j, 3] _lowercase : Optional[int] = bbox[i, j, 1] _lowercase : str = t if bbox[i, j, 2] < bbox[i, j, 0]: _lowercase : Dict = bbox[i, j, 2] _lowercase : Any = bbox[i, j, 0] _lowercase : int = t _lowercase : int = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowercase : int = None if self.use_input_mask: _lowercase : Union[str, Any] = random_attention_mask([self.batch_size, self.text_seq_length] ) _lowercase : int = None if self.use_token_type_ids: _lowercase : List[str] = ids_tensor([self.batch_size, self.text_seq_length] , self.type_vocab_size ) _lowercase : List[Any] = None _lowercase : Union[str, Any] = None if self.use_labels: _lowercase : Dict = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowercase : Dict = ids_tensor([self.batch_size, self.text_seq_length] , self.num_labels ) _lowercase : List[Any] = LayoutLMvaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , coordinate_size=self.coordinate_size , shape_size=self.shape_size , input_size=self.image_size , patch_size=self.patch_size , ) return config, input_ids, bbox, pixel_values, token_type_ids, input_mask, sequence_labels, token_labels def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): _lowercase : Optional[int] = LayoutLMvaModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() # text + image _lowercase : str = model(_lowerCAmelCase , pixel_values=_lowerCAmelCase ) _lowercase : Optional[Any] = model( _lowerCAmelCase , bbox=_lowerCAmelCase , pixel_values=_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) _lowercase : List[Any] = model(_lowerCAmelCase , bbox=_lowerCAmelCase , pixel_values=_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) _lowercase : List[str] = model(_lowerCAmelCase , bbox=_lowerCAmelCase , pixel_values=_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # text only _lowercase : List[str] = model(_lowerCAmelCase ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.text_seq_length, self.hidden_size) ) # image only _lowercase : str = model(pixel_values=_lowerCAmelCase ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.image_seq_length, self.hidden_size) ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): _lowercase : Any = self.num_labels _lowercase : str = LayoutLMvaForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() _lowercase : Any = model( _lowerCAmelCase , bbox=_lowerCAmelCase , pixel_values=_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): _lowercase : List[Any] = self.num_labels _lowercase : Optional[Any] = LayoutLMvaForTokenClassification(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() _lowercase : Tuple = model( _lowerCAmelCase , bbox=_lowerCAmelCase , pixel_values=_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.text_seq_length, self.num_labels) ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): _lowercase : str = LayoutLMvaForQuestionAnswering(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() _lowercase : Union[str, Any] = model( _lowerCAmelCase , bbox=_lowerCAmelCase , pixel_values=_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , start_positions=_lowerCAmelCase , end_positions=_lowerCAmelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __a ( self ): _lowercase : List[Any] = self.prepare_config_and_inputs() ( ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ) : Union[str, Any] = config_and_inputs _lowercase : Dict = { 'input_ids': input_ids, 'bbox': bbox, 'pixel_values': pixel_values, 'token_type_ids': token_type_ids, 'attention_mask': input_mask, } return config, inputs_dict @require_torch class lowerCAmelCase_ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCamelCase : List[Any] = False _UpperCamelCase : str = False _UpperCamelCase : Any = False _UpperCamelCase : Dict = ( ( LayoutLMvaModel, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaForQuestionAnswering, ) if is_torch_available() else () ) _UpperCamelCase : Any = ( {"document-question-answering": LayoutLMvaForQuestionAnswering, "feature-extraction": LayoutLMvaModel} if is_torch_available() else {} ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): # `DocumentQuestionAnsweringPipeline` is expected to work with this model, but it combines the text and visual # embedding along the sequence dimension (dim 1), which causes an error during post-processing as `p_mask` has # the sequence dimension of the text embedding only. # (see the line `embedding_output = torch.cat([embedding_output, visual_embeddings], dim=1)`) return True def __a ( self ): _lowercase : List[Any] = LayoutLMvaModelTester(self ) _lowercase : Optional[int] = ConfigTester(self , config_class=_lowerCAmelCase , hidden_size=3_7 ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase=False ): _lowercase : Optional[int] = copy.deepcopy(_lowerCAmelCase ) if model_class in get_values(_lowerCAmelCase ): _lowercase : Tuple = { k: v.unsqueeze(1 ).expand(-1 , self.model_tester.num_choices , -1 ).contiguous() if isinstance(_lowerCAmelCase , torch.Tensor ) and v.ndim > 1 else v for k, v in inputs_dict.items() } if return_labels: if model_class in get_values(_lowerCAmelCase ): _lowercase : Tuple = torch.ones(self.model_tester.batch_size , dtype=torch.long , device=_lowerCAmelCase ) elif model_class in get_values(_lowerCAmelCase ): _lowercase : str = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_lowerCAmelCase ) _lowercase : Any = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_lowerCAmelCase ) elif model_class in [ *get_values(_lowerCAmelCase ), ]: _lowercase : Any = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_lowerCAmelCase ) elif model_class in [ *get_values(_lowerCAmelCase ), ]: _lowercase : List[Any] = torch.zeros( (self.model_tester.batch_size, self.model_tester.text_seq_length) , dtype=torch.long , device=_lowerCAmelCase , ) return inputs_dict def __a ( self ): self.config_tester.run_common_tests() def __a ( self ): _lowercase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def __a ( self ): _lowercase : Tuple = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _lowercase : Union[str, Any] = type self.model_tester.create_and_check_model(*_lowerCAmelCase ) def __a ( self ): _lowercase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_lowerCAmelCase ) def __a ( self ): _lowercase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_lowerCAmelCase ) def __a ( self ): _lowercase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_lowerCAmelCase ) @slow def __a ( self ): for model_name in LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowercase : Tuple = LayoutLMvaModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def __magic_name__ ( ) -> Any: _lowercase : Any = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch class lowerCAmelCase_ ( unittest.TestCase ): @cached_property def __a ( self ): return LayoutLMvaImageProcessor(apply_ocr=_lowerCAmelCase ) if is_vision_available() else None @slow def __a ( self ): _lowercase : Tuple = LayoutLMvaModel.from_pretrained('microsoft/layoutlmv3-base' ).to(_lowerCAmelCase ) _lowercase : List[Any] = self.default_image_processor _lowercase : List[str] = prepare_img() _lowercase : Union[str, Any] = image_processor(images=_lowerCAmelCase , return_tensors='pt' ).pixel_values.to(_lowerCAmelCase ) _lowercase : Tuple = torch.tensor([[1, 2]] ) _lowercase : Any = torch.tensor([[1, 2, 3, 4], [5, 6, 7, 8]] ).unsqueeze(0 ) # forward pass _lowercase : int = model( input_ids=input_ids.to(_lowerCAmelCase ) , bbox=bbox.to(_lowerCAmelCase ) , pixel_values=pixel_values.to(_lowerCAmelCase ) , ) # verify the logits _lowercase : Union[str, Any] = torch.Size((1, 1_9_9, 7_6_8) ) self.assertEqual(outputs.last_hidden_state.shape , _lowerCAmelCase ) _lowercase : int = torch.tensor( [[-0.05_29, 0.36_18, 0.16_32], [-0.15_87, -0.16_67, -0.04_00], [-0.15_57, -0.16_71, -0.05_05]] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , _lowerCAmelCase , atol=1E-4 ) )
66
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): _lowercase : List[Any] = torch.nn.Linear(1_0 , 1_0 ) _lowercase : Any = torch.optim.SGD(model.parameters() , 0.1 ) _lowercase : str = Accelerator() _lowercase : Any = accelerator.prepare(_lowerCAmelCase ) try: pickle.loads(pickle.dumps(_lowerCAmelCase ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
66
1
from math import ceil, sqrt def __magic_name__ ( SCREAMING_SNAKE_CASE = 1_000_000 ) -> int: _lowercase : str = 0 for outer_width in range(3 , (limit // 4) + 2 ): if outer_width**2 > limit: _lowercase : Any = max(ceil(sqrt(outer_width**2 - limit ) ) , 1 ) else: _lowercase : Optional[Any] = 1 if (outer_width - hole_width_lower_bound) % 2: hole_width_lower_bound += 1 answer += (outer_width - hole_width_lower_bound - 2) // 2 + 1 return answer if __name__ == "__main__": print(f'''{solution() = }''')
66
import requests from bsa import BeautifulSoup def __magic_name__ ( SCREAMING_SNAKE_CASE = "AAPL" ) -> str: _lowercase : str = F"""https://in.finance.yahoo.com/quote/{symbol}?s={symbol}""" _lowercase : int = BeautifulSoup(requests.get(SCREAMING_SNAKE_CASE ).text , 'html.parser' ) _lowercase : List[str] = 'My(6px) Pos(r) smartphone_Mt(6px)' return soup.find('div' , class_=class_ ).find('span' ).text if __name__ == "__main__": for symbol in "AAPL AMZN IBM GOOG MSFT ORCL".split(): print(f'''Current {symbol:<4} stock price is {stock_price(symbol):>8}''')
66
1
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> float: return price * (1 + tax_rate) if __name__ == "__main__": print(f'''{price_plus_tax(100, 0.25) = }''') print(f'''{price_plus_tax(1_25.50, 0.05) = }''')
66
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = { "configuration_poolformer": [ "POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "PoolFormerConfig", "PoolFormerOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["PoolFormerFeatureExtractor"] UpperCamelCase = ["PoolFormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "PoolFormerForImageClassification", "PoolFormerModel", "PoolFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure)
66
1
import unittest from transformers import JukeboxTokenizer from transformers.testing_utils import require_torch class lowerCAmelCase_ ( unittest.TestCase ): _UpperCamelCase : Optional[Any] = JukeboxTokenizer _UpperCamelCase : Tuple = { "artist": "Zac Brown Band", "genres": "Country", "lyrics": "I met a traveller from an antique land,\n Who said \"Two vast and trunkless legs of stone\n Stand in the desert. . . . Near them, on the sand,\n Half sunk a shattered visage lies, whose frown,\n And wrinkled lip, and sneer of cold command,\n Tell that its sculptor well those passions read\n Which yet survive, stamped on these lifeless things,\n The hand that mocked them, and the heart that fed;\n And on the pedestal, these words appear:\n My name is Ozymandias, King of Kings;\n Look on my Works, ye Mighty, and despair!\n Nothing beside remains. Round the decay\n Of that colossal Wreck, boundless and bare\n The lone and level sands stretch far away\n ", } @require_torch def __a ( self ): import torch _lowercase : List[Any] = JukeboxTokenizer.from_pretrained('openai/jukebox-1b-lyrics' ) _lowercase : Optional[Any] = tokenizer(**self.metas )['input_ids'] # fmt: off _lowercase : List[str] = [ torch.tensor([[ 0, 0, 0, 7_1_6_9, 5_0_7, 9, 7_6, 3_9, 3_1, 4_6, 7_6, 2_7, 7_6, 4_6, 4_4, 2_7, 4_8, 3_1, 3_8, 3_8, 3_1, 4_4, 7_6, 3_2, 4_4, 4_1, 3_9, 7_6, 2_7, 4_0, 7_6, 2_7, 4_0, 4_6, 3_5, 4_3, 4_7, 3_1, 7_6, 3_8, 2_7, 4_0, 3_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_3, 3_4, 4_1, 7_6, 4_5, 2_7, 3_5, 3_0, 7_6, 7_1, 2_0, 4_9, 4_1, 7_6, 4_8, 2_7, 4_5, 4_6, 7_6, 2_7, 4_0, 3_0, 7_6, 4_6, 4_4, 4_7, 4_0, 3_7, 3_8, 3_1, 4_5, 4_5, 7_6, 3_8, 3_1, 3_3, 4_5, 7_6, 4_1, 3_2, 7_6, 4_5, 4_6, 4_1, 4_0, 3_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_9, 4_6, 2_7, 4_0, 3_0, 7_6, 3_5, 4_0, 7_6, 4_6, 3_4, 3_1, 7_6, 3_0, 3_1, 4_5, 3_1, 4_4, 4_6, 6_3, 7_6, 6_3, 7_6, 6_3, 7_6, 6_3, 7_6, 1_4, 3_1, 2_7, 4_4, 7_6, 4_6, 3_4, 3_1, 3_9, 6_4, 7_6, 4_1, 4_0, 7_6, 4_6, 3_4, 3_1, 7_6, 4_5, 2_7, 4_0, 3_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 8, 2_7, 3_8, 3_2, 7_6, 4_5, 4_7, 4_0, 3_7, 7_6, 2_7, 7_6, 4_5, 3_4, 2_7, 4_6, 4_6, 3_1, 4_4, 3_1, 3_0, 7_6, 4_8, 3_5, 4_5, 2_7, 3_3, 3_1, 7_6, 3_8, 3_5, 3_1, 4_5, 6_4, 7_6, 4_9, 3_4, 4_1, 4_5, 3_1, 7_6, 3_2, 4_4, 4_1, 4_9, 4_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1, 4_0, 3_0, 7_6, 4_9, 4_4, 3_5, 4_0, 3_7, 3_8, 3_1, 3_0, 7_6, 3_8, 3_5, 4_2, 6_4, 7_6, 2_7, 4_0, 3_0, 7_6, 4_5, 4_0, 3_1, 3_1, 4_4, 7_6, 4_1, 3_2, 7_6, 2_9, 4_1, 3_8, 3_0, 7_6, 2_9, 4_1, 3_9, 3_9, 2_7, 4_0, 3_0, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_0, 3_1, 3_8, 3_8, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 3_5, 4_6, 4_5, 7_6, 4_5, 2_9, 4_7, 3_8, 4_2, 4_6, 4_1, 4_4, 7_6, 4_9, 3_1, 3_8, 3_8, 7_6, 4_6, 3_4, 4_1, 4_5, 3_1, 7_6, 4_2, 2_7, 4_5, 4_5, 3_5, 4_1, 4_0, 4_5, 7_6, 4_4, 3_1, 2_7, 3_0, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_3, 3_4, 3_5, 2_9, 3_4, 7_6, 5_1, 3_1, 4_6, 7_6, 4_5, 4_7, 4_4, 4_8, 3_5, 4_8, 3_1, 6_4, 7_6, 4_5, 4_6, 2_7, 3_9, 4_2, 3_1, 3_0, 7_6, 4_1, 4_0, 7_6, 4_6, 3_4, 3_1, 4_5, 3_1, 7_6, 3_8, 3_5, 3_2, 3_1, 3_8, 3_1, 4_5, 4_5, 7_6, 4_6, 3_4, 3_5, 4_0, 3_3, 4_5, 6_4, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_0, 3_4, 3_1, 7_6, 3_4, 2_7, 4_0, 3_0, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 3_9, 4_1, 2_9, 3_7, 3_1, 3_0, 7_6, 4_6, 3_4, 3_1, 3_9, 6_4, 7_6, 2_7, 4_0, 3_0, 7_6, 4_6, 3_4, 3_1, 7_6, 3_4, 3_1, 2_7, 4_4, 4_6, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 3_2, 3_1, 3_0, 6_6, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1, 4_0, 3_0, 7_6, 4_1, 4_0, 7_6, 4_6, 3_4, 3_1, 7_6, 4_2, 3_1, 3_0, 3_1, 4_5, 4_6, 2_7, 3_8, 6_4, 7_6, 4_6, 3_4, 3_1, 4_5, 3_1, 7_6, 4_9, 4_1, 4_4, 3_0, 4_5, 7_6, 2_7, 4_2, 4_2, 3_1, 2_7, 4_4, 6_5, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_3, 5_1, 7_6, 4_0, 2_7, 3_9, 3_1, 7_6, 3_5, 4_5, 7_6, 1_5, 5_2, 5_1, 3_9, 2_7, 4_0, 3_0, 3_5, 2_7, 4_5, 6_4, 7_6, 1_1, 3_5, 4_0, 3_3, 7_6, 4_1, 3_2, 7_6, 1_1, 3_5, 4_0, 3_3, 4_5, 6_6, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_2, 4_1, 4_1, 3_7, 7_6, 4_1, 4_0, 7_6, 3_9, 5_1, 7_6, 2_3, 4_1, 4_4, 3_7, 4_5, 6_4, 7_6, 5_1, 3_1, 7_6, 1_3, 3_5, 3_3, 3_4, 4_6, 5_1, 6_4, 7_6, 2_7, 4_0, 3_0, 7_6, 3_0, 3_1, 4_5, 4_2, 2_7, 3_5, 4_4, 6_7, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_4, 4_1, 4_6, 3_4, 3_5, 4_0, 3_3, 7_6, 2_8, 3_1, 4_5, 3_5, 3_0, 3_1, 7_6, 4_4, 3_1, 3_9, 2_7, 3_5, 4_0, 4_5, 6_3, 7_6, 1_8, 4_1, 4_7, 4_0, 3_0, 7_6, 4_6, 3_4, 3_1, 7_6, 3_0, 3_1, 2_9, 2_7, 5_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 1_5, 3_2, 7_6, 4_6, 3_4, 2_7, 4_6, 7_6, 2_9, 4_1, 3_8, 4_1, 4_5, 4_5, 2_7, 3_8, 7_6, 2_3, 4_4, 3_1, 2_9, 3_7, 6_4, 7_6, 2_8, 4_1, 4_7, 4_0, 3_0, 3_8, 3_1, 4_5, 4_5, 7_6, 2_7, 4_0, 3_0, 7_6, 2_8, 2_7, 4_4, 3_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 2_0, 3_4, 3_1, 7_6, 3_8, 4_1, 4_0, 3_1, 7_6, 2_7, 4_0, 3_0, 7_6, 3_8, 3_1, 4_8, 3_1, 3_8, 7_6, 4_5, 2_7, 4_0, 3_0, 4_5, 7_6, 4_5, 4_6, 4_4, 3_1, 4_6, 2_9, 3_4, 7_6, 3_2, 2_7, 4_4, 7_6, 2_7, 4_9, 2_7, 5_1, 7_8, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6, 7_6]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) ) @require_torch def __a ( self ): import torch _lowercase : Optional[int] = JukeboxTokenizer.from_pretrained('openai/jukebox-5b-lyrics' ) _lowercase : Tuple = tokenizer(**self.metas )['input_ids'] # fmt: off _lowercase : Optional[Any] = [ torch.tensor([[ 0, 0, 0, 1_0_6_9, 1_1, -1, -1, -1, -1, 9, 7_7, 3_9, 3_1, 4_6, 7_7, 2_7, 7_7, 4_6, 4_4, 2_7, 4_8, 3_1, 3_8, 3_8, 3_1, 4_4, 7_7, 3_2, 4_4, 4_1, 3_9, 7_7, 2_7, 4_0, 7_7, 2_7, 4_0, 4_6, 3_5, 4_3, 4_7, 3_1, 7_7, 3_8, 2_7, 4_0, 3_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_3, 3_4, 4_1, 7_7, 4_5, 2_7, 3_5, 3_0, 7_7, 7_2, 2_0, 4_9, 4_1, 7_7, 4_8, 2_7, 4_5, 4_6, 7_7, 2_7, 4_0, 3_0, 7_7, 4_6, 4_4, 4_7, 4_0, 3_7, 3_8, 3_1, 4_5, 4_5, 7_7, 3_8, 3_1, 3_3, 4_5, 7_7, 4_1, 3_2, 7_7, 4_5, 4_6, 4_1, 4_0, 3_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_9, 4_6, 2_7, 4_0, 3_0, 7_7, 3_5, 4_0, 7_7, 4_6, 3_4, 3_1, 7_7, 3_0, 3_1, 4_5, 3_1, 4_4, 4_6, 6_3, 7_7, 6_3, 7_7, 6_3, 7_7, 6_3, 7_7, 1_4, 3_1, 2_7, 4_4, 7_7, 4_6, 3_4, 3_1, 3_9, 6_4, 7_7, 4_1, 4_0, 7_7, 4_6, 3_4, 3_1, 7_7, 4_5, 2_7, 4_0, 3_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 8, 2_7, 3_8, 3_2, 7_7, 4_5, 4_7, 4_0, 3_7, 7_7, 2_7, 7_7, 4_5, 3_4, 2_7, 4_6, 4_6, 3_1, 4_4, 3_1, 3_0, 7_7, 4_8, 3_5, 4_5, 2_7, 3_3, 3_1, 7_7, 3_8, 3_5, 3_1, 4_5, 6_4, 7_7, 4_9, 3_4, 4_1, 4_5, 3_1, 7_7, 3_2, 4_4, 4_1, 4_9, 4_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1, 4_0, 3_0, 7_7, 4_9, 4_4, 3_5, 4_0, 3_7, 3_8, 3_1, 3_0, 7_7, 3_8, 3_5, 4_2, 6_4, 7_7, 2_7, 4_0, 3_0, 7_7, 4_5, 4_0, 3_1, 3_1, 4_4, 7_7, 4_1, 3_2, 7_7, 2_9, 4_1, 3_8, 3_0, 7_7, 2_9, 4_1, 3_9, 3_9, 2_7, 4_0, 3_0, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_0, 3_1, 3_8, 3_8, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 3_5, 4_6, 4_5, 7_7, 4_5, 2_9, 4_7, 3_8, 4_2, 4_6, 4_1, 4_4, 7_7, 4_9, 3_1, 3_8, 3_8, 7_7, 4_6, 3_4, 4_1, 4_5, 3_1, 7_7, 4_2, 2_7, 4_5, 4_5, 3_5, 4_1, 4_0, 4_5, 7_7, 4_4, 3_1, 2_7, 3_0, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_3, 3_4, 3_5, 2_9, 3_4, 7_7, 5_1, 3_1, 4_6, 7_7, 4_5, 4_7, 4_4, 4_8, 3_5, 4_8, 3_1, 6_4, 7_7, 4_5, 4_6, 2_7, 3_9, 4_2, 3_1, 3_0, 7_7, 4_1, 4_0, 7_7, 4_6, 3_4, 3_1, 4_5, 3_1, 7_7, 3_8, 3_5, 3_2, 3_1, 3_8, 3_1, 4_5, 4_5, 7_7, 4_6, 3_4, 3_5, 4_0, 3_3, 4_5, 6_4, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_0, 3_4, 3_1, 7_7, 3_4, 2_7, 4_0, 3_0, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 3_9, 4_1, 2_9, 3_7, 3_1, 3_0, 7_7, 4_6, 3_4, 3_1, 3_9, 6_4, 7_7, 2_7, 4_0, 3_0, 7_7, 4_6, 3_4, 3_1, 7_7, 3_4, 3_1, 2_7, 4_4, 4_6, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 3_2, 3_1, 3_0, 6_6, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1, 4_0, 3_0, 7_7, 4_1, 4_0, 7_7, 4_6, 3_4, 3_1, 7_7, 4_2, 3_1, 3_0, 3_1, 4_5, 4_6, 2_7, 3_8, 6_4, 7_7, 4_6, 3_4, 3_1, 4_5, 3_1, 7_7, 4_9, 4_1, 4_4, 3_0, 4_5, 7_7, 2_7, 4_2, 4_2, 3_1, 2_7, 4_4, 6_5, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_3, 5_1, 7_7, 4_0, 2_7, 3_9, 3_1, 7_7, 3_5, 4_5, 7_7, 1_5, 5_2, 5_1, 3_9, 2_7, 4_0, 3_0, 3_5, 2_7, 4_5, 6_4, 7_7, 1_1, 3_5, 4_0, 3_3, 7_7, 4_1, 3_2, 7_7, 1_1, 3_5, 4_0, 3_3, 4_5, 6_6, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_2, 4_1, 4_1, 3_7, 7_7, 4_1, 4_0, 7_7, 3_9, 5_1, 7_7, 2_3, 4_1, 4_4, 3_7, 4_5, 6_4, 7_7, 5_1, 3_1, 7_7, 1_3, 3_5, 3_3, 3_4, 4_6, 5_1, 6_4, 7_7, 2_7, 4_0, 3_0, 7_7, 3_0, 3_1, 4_5, 4_2, 2_7, 3_5, 4_4, 6_7, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_4, 4_1, 4_6, 3_4, 3_5, 4_0, 3_3, 7_7, 2_8, 3_1, 4_5, 3_5, 3_0, 3_1, 7_7, 4_4, 3_1, 3_9, 2_7, 3_5, 4_0, 4_5, 6_3, 7_7, 1_8, 4_1, 4_7, 4_0, 3_0, 7_7, 4_6, 3_4, 3_1, 7_7, 3_0, 3_1, 2_9, 2_7, 5_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 1_5, 3_2, 7_7, 4_6, 3_4, 2_7, 4_6, 7_7, 2_9, 4_1, 3_8, 4_1, 4_5, 4_5, 2_7, 3_8, 7_7, 2_3, 4_4, 3_1, 2_9, 3_7, 6_4, 7_7, 2_8, 4_1, 4_7, 4_0, 3_0, 3_8, 3_1, 4_5, 4_5, 7_7, 2_7, 4_0, 3_0, 7_7, 2_8, 2_7, 4_4, 3_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 2_0, 3_4, 3_1, 7_7, 3_8, 4_1, 4_0, 3_1, 7_7, 2_7, 4_0, 3_0, 7_7, 3_8, 3_1, 4_8, 3_1, 3_8, 7_7, 4_5, 2_7, 4_0, 3_0, 4_5, 7_7, 4_5, 4_6, 4_4, 3_1, 4_6, 2_9, 3_4, 7_7, 3_2, 2_7, 4_4, 7_7, 2_7, 4_9, 2_7, 5_1, 7_9, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7, 7_7]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1, -1, -1, -1, -1]] ), torch.tensor([[0, 0, 0, 1_0_6_9, 1_1, -1, -1, -1, -1]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) )
66
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : Dict = LayoutLMTokenizer _UpperCamelCase : Union[str, Any] = LayoutLMTokenizerFast _UpperCamelCase : int = True _UpperCamelCase : Tuple = True def __a ( self ): super().setUp() _lowercase : Union[str, Any] = [ '[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] _lowercase : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def __a ( self , **_lowerCAmelCase ): return LayoutLMTokenizer.from_pretrained(self.tmpdirname , **_lowerCAmelCase ) def __a ( self , _lowerCAmelCase ): _lowercase : str = 'UNwant\u00E9d,running' _lowercase : List[Any] = 'unwanted, running' return input_text, output_text def __a ( self ): _lowercase : Dict = self.tokenizer_class(self.vocab_file ) _lowercase : Dict = tokenizer.tokenize('UNwant\u00E9d,running' ) self.assertListEqual(_lowerCAmelCase , ['un', '##want', '##ed', ',', 'runn', '##ing'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [7, 4, 5, 1_0, 8, 9] ) def __a ( self ): pass
66
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = { "configuration_chinese_clip": [ "CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "ChineseCLIPConfig", "ChineseCLIPOnnxConfig", "ChineseCLIPTextConfig", "ChineseCLIPVisionConfig", ], "processing_chinese_clip": ["ChineseCLIPProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["ChineseCLIPFeatureExtractor"] UpperCamelCase = ["ChineseCLIPImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "ChineseCLIPModel", "ChineseCLIPPreTrainedModel", "ChineseCLIPTextModel", "ChineseCLIPVisionModel", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : str = ShapEPipeline _UpperCamelCase : Any = ["prompt"] _UpperCamelCase : int = ["prompt"] _UpperCamelCase : Union[str, Any] = [ "num_images_per_prompt", "num_inference_steps", "generator", "latents", "guidance_scale", "frame_size", "output_type", "return_dict", ] _UpperCamelCase : Optional[Any] = False @property def __a ( self ): return 3_2 @property def __a ( self ): return 3_2 @property def __a ( self ): return self.time_input_dim * 4 @property def __a ( self ): return 8 @property def __a ( self ): _lowercase : Any = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) return tokenizer @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) return CLIPTextModelWithProjection(_lowerCAmelCase ) @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = { 'num_attention_heads': 2, 'attention_head_dim': 1_6, 'embedding_dim': self.time_input_dim, 'num_embeddings': 3_2, 'embedding_proj_dim': self.text_embedder_hidden_size, 'time_embed_dim': self.time_embed_dim, 'num_layers': 1, 'clip_embed_dim': self.time_input_dim * 2, 'additional_embeddings': 0, 'time_embed_act_fn': 'gelu', 'norm_in_type': 'layer', 'encoder_hid_proj_type': None, 'added_emb_type': None, } _lowercase : Optional[Any] = PriorTransformer(**_lowerCAmelCase ) return model @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = { 'param_shapes': ( (self.renderer_dim, 9_3), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), 'd_latent': self.time_input_dim, 'd_hidden': self.renderer_dim, 'n_output': 1_2, 'background': ( 0.1, 0.1, 0.1, ), } _lowercase : List[Any] = ShapERenderer(**_lowerCAmelCase ) return model def __a ( self ): _lowercase : Optional[Any] = self.dummy_prior _lowercase : Dict = self.dummy_text_encoder _lowercase : List[str] = self.dummy_tokenizer _lowercase : Union[str, Any] = self.dummy_renderer _lowercase : List[str] = HeunDiscreteScheduler( beta_schedule='exp' , num_train_timesteps=1_0_2_4 , prediction_type='sample' , use_karras_sigmas=_lowerCAmelCase , clip_sample=_lowerCAmelCase , clip_sample_range=1.0 , ) _lowercase : List[str] = { 'prior': prior, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'renderer': renderer, 'scheduler': scheduler, } return components def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : Optional[Any] = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Optional[int] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : List[Any] = { 'prompt': 'horse', 'generator': generator, 'num_inference_steps': 1, 'frame_size': 3_2, 'output_type': 'np', } return inputs def __a ( self ): _lowercase : Optional[int] = 'cpu' _lowercase : List[Any] = self.get_dummy_components() _lowercase : Tuple = self.pipeline_class(**_lowerCAmelCase ) _lowercase : Union[str, Any] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Union[str, Any] = pipe(**self.get_dummy_inputs(_lowerCAmelCase ) ) _lowercase : str = output.images[0] _lowercase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (2_0, 3_2, 3_2, 3) _lowercase : str = np.array( [ 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __a ( self ): # NOTE: Larger batch sizes cause this test to timeout, only test on smaller batches self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def __a ( self ): _lowercase : List[Any] = torch_device == 'cpu' _lowercase : Any = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=_lowerCAmelCase , relax_max_difference=_lowerCAmelCase , ) def __a ( self ): _lowercase : Union[str, Any] = self.get_dummy_components() _lowercase : Optional[int] = self.pipeline_class(**_lowerCAmelCase ) _lowercase : Any = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : str = 1 _lowercase : Optional[int] = 2 _lowercase : List[str] = self.get_dummy_inputs(_lowerCAmelCase ) for key in inputs.keys(): if key in self.batch_params: _lowercase : int = batch_size * [inputs[key]] _lowercase : Optional[int] = pipe(**_lowerCAmelCase , num_images_per_prompt=_lowerCAmelCase )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ): _lowercase : List[str] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/shap_e/test_shap_e_np_out.npy' ) _lowercase : Any = ShapEPipeline.from_pretrained('openai/shap-e' ) _lowercase : List[str] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Tuple = torch.Generator(device=_lowerCAmelCase ).manual_seed(0 ) _lowercase : int = pipe( 'a shark' , generator=_lowerCAmelCase , guidance_scale=15.0 , num_inference_steps=6_4 , frame_size=6_4 , output_type='np' , ).images[0] assert images.shape == (2_0, 6_4, 6_4, 3) assert_mean_pixel_difference(_lowerCAmelCase , _lowerCAmelCase )
66
1
import torch from diffusers import DiffusionPipeline class lowerCAmelCase_ ( __snake_case ): def __init__( self , _lowerCAmelCase , _lowerCAmelCase ): super().__init__() self.register_modules(unet=_lowerCAmelCase , scheduler=_lowerCAmelCase ) def __call__( self ): _lowercase : str = torch.randn( (1, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , ) _lowercase : Union[str, Any] = 1 _lowercase : str = self.unet(_lowerCAmelCase , _lowerCAmelCase ).sample _lowercase : Optional[int] = self.scheduler.step(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ).prev_sample _lowercase : int = scheduler_output - scheduler_output + torch.ones_like(_lowerCAmelCase ) return result
66
import sys UpperCamelCase = ( "73167176531330624919225119674426574742355349194934" "96983520312774506326239578318016984801869478851843" "85861560789112949495459501737958331952853208805511" "12540698747158523863050715693290963295227443043557" "66896648950445244523161731856403098711121722383113" "62229893423380308135336276614282806444486645238749" "30358907296290491560440772390713810515859307960866" "70172427121883998797908792274921901699720888093776" "65727333001053367881220235421809751254540594752243" "52584907711670556013604839586446706324415722155397" "53697817977846174064955149290862569321978468622482" "83972241375657056057490261407972968652414535100474" "82166370484403199890008895243450658541227588666881" "16427171479924442928230863465674813919123162824586" "17866458359124566529476545682848912883142607690042" "24219022671055626321111109370544217506941658960408" "07198403850962455444362981230987879927244284909188" "84580156166097919133875499200524063689912560717606" "05886116467109405077541002256983155200055935729725" "71636269561882670428252483600823257530420752963450" ) def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : List[Any] = 1 for digit in s: product *= int(SCREAMING_SNAKE_CASE ) return product def __magic_name__ ( SCREAMING_SNAKE_CASE = N ) -> int: _lowercase : Dict = -sys.maxsize - 1 _lowercase : Tuple = n[:13] _lowercase : List[Any] = 13 while cur_index < len(SCREAMING_SNAKE_CASE ) - 13: if int(n[cur_index] ) >= int(substr[0] ): _lowercase : List[str] = substr[1:] + n[cur_index] cur_index += 1 else: _lowercase : str = max(SCREAMING_SNAKE_CASE , str_eval(SCREAMING_SNAKE_CASE ) ) _lowercase : Dict = n[cur_index : cur_index + 13] cur_index += 13 return largest_product if __name__ == "__main__": print(f'''{solution() = }''')
66
1
def __magic_name__ ( SCREAMING_SNAKE_CASE = 200 ) -> int: _lowercase : Optional[Any] = [1, 2, 5, 10, 20, 50, 100, 200] _lowercase : List[str] = [0] * (pence + 1) _lowercase : Any = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(SCREAMING_SNAKE_CASE , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 73_682
66
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available UpperCamelCase = { "configuration_gpt_neo": ["GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTNeoConfig", "GPTNeoOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTNeoForCausalLM", "GPTNeoForQuestionAnswering", "GPTNeoForSequenceClassification", "GPTNeoForTokenClassification", "GPTNeoModel", "GPTNeoPreTrainedModel", "load_tf_weights_in_gpt_neo", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "FlaxGPTNeoForCausalLM", "FlaxGPTNeoModel", "FlaxGPTNeoPreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_neo import GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoConfig, GPTNeoOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neo import ( GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoForCausalLM, GPTNeoForQuestionAnswering, GPTNeoForSequenceClassification, GPTNeoForTokenClassification, GPTNeoModel, GPTNeoPreTrainedModel, load_tf_weights_in_gpt_neo, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_gpt_neo import FlaxGPTNeoForCausalLM, FlaxGPTNeoModel, FlaxGPTNeoPreTrainedModel else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
1
import math def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> list[int]: _lowercase : int = [] _lowercase : Optional[Any] = 2 _lowercase : Union[str, Any] = int(math.sqrt(SCREAMING_SNAKE_CASE ) ) # Size of every segment _lowercase : Union[str, Any] = [True] * (end + 1) _lowercase : List[str] = [] while start <= end: if temp[start] is True: in_prime.append(SCREAMING_SNAKE_CASE ) for i in range(start * start , end + 1 , SCREAMING_SNAKE_CASE ): _lowercase : Tuple = False start += 1 prime += in_prime _lowercase : int = end + 1 _lowercase : List[str] = min(2 * end , SCREAMING_SNAKE_CASE ) while low <= n: _lowercase : str = [True] * (high - low + 1) for each in in_prime: _lowercase : Dict = math.floor(low / each ) * each if t < low: t += each for j in range(SCREAMING_SNAKE_CASE , high + 1 , SCREAMING_SNAKE_CASE ): _lowercase : Optional[Any] = False for j in range(len(SCREAMING_SNAKE_CASE ) ): if temp[j] is True: prime.append(j + low ) _lowercase : Tuple = high + 1 _lowercase : Any = min(high + end , SCREAMING_SNAKE_CASE ) return prime print(sieve(10**6))
66
from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : str = ["image_processor", "tokenizer"] _UpperCamelCase : Union[str, Any] = "AutoImageProcessor" _UpperCamelCase : Union[str, Any] = "AutoTokenizer" def __init__( self , _lowerCAmelCase , _lowerCAmelCase ): super().__init__(_lowerCAmelCase , _lowerCAmelCase ) _lowercase : Union[str, Any] = self.image_processor def __call__( self , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=None , **_lowerCAmelCase ): if text is None and images is None: raise ValueError('You have to specify either text or images. Both cannot be none.' ) if text is not None: _lowercase : Dict = self.tokenizer(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if images is not None: _lowercase : Union[str, Any] = self.image_processor(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if text is not None and images is not None: _lowercase : Optional[Any] = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_lowerCAmelCase ) , tensor_type=_lowerCAmelCase ) def __a ( self , *_lowerCAmelCase , **_lowerCAmelCase ): return self.tokenizer.batch_decode(*_lowerCAmelCase , **_lowerCAmelCase ) def __a ( self , *_lowerCAmelCase , **_lowerCAmelCase ): return self.tokenizer.decode(*_lowerCAmelCase , **_lowerCAmelCase ) @property def __a ( self ): return ["input_ids", "attention_mask", "pixel_values"]
66
1
from collections import deque from math import floor from random import random from time import time class lowerCAmelCase_ : def __init__( self ): _lowercase : Optional[int] = {} def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase=1 ): if self.graph.get(_lowerCAmelCase ): if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: _lowercase : Tuple = [[w, v]] if not self.graph.get(_lowerCAmelCase ): _lowercase : Optional[Any] = [] def __a ( self ): return list(self.graph ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase ): if self.graph.get(_lowerCAmelCase ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(_lowerCAmelCase ) def __a ( self , _lowerCAmelCase=-2 , _lowerCAmelCase=-1 ): if s == d: return [] _lowercase : str = [] _lowercase : Union[str, Any] = [] if s == -2: _lowercase : str = list(self.graph )[0] stack.append(_lowerCAmelCase ) visited.append(_lowerCAmelCase ) _lowercase : Optional[int] = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: _lowercase : Optional[int] = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(_lowerCAmelCase ) return visited else: stack.append(node[1] ) visited.append(node[1] ) _lowercase : Union[str, Any] = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(_lowerCAmelCase ) != 0: _lowercase : int = stack[len(_lowerCAmelCase ) - 1] else: _lowercase : Optional[int] = ss # check if se have reached the starting point if len(_lowerCAmelCase ) == 0: return visited def __a ( self , _lowerCAmelCase=-1 ): if c == -1: _lowercase : Dict = floor(random() * 1_0_0_0_0 ) + 1_0 for i in range(_lowerCAmelCase ): # every vertex has max 100 edges for _ in range(floor(random() * 1_0_2 ) + 1 ): _lowercase : str = floor(random() * c ) + 1 if n != i: self.add_pair(_lowerCAmelCase , _lowerCAmelCase , 1 ) def __a ( self , _lowerCAmelCase=-2 ): _lowercase : str = deque() _lowercase : Optional[Any] = [] if s == -2: _lowercase : List[Any] = list(self.graph )[0] d.append(_lowerCAmelCase ) visited.append(_lowerCAmelCase ) while d: _lowercase : Tuple = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def __a ( self , _lowerCAmelCase ): _lowercase : str = 0 for x in self.graph: for y in self.graph[x]: if y[1] == u: count += 1 return count def __a ( self , _lowerCAmelCase ): return len(self.graph[u] ) def __a ( self , _lowerCAmelCase=-2 ): _lowercase : Dict = [] _lowercase : Tuple = [] if s == -2: _lowercase : Union[str, Any] = list(self.graph )[0] stack.append(_lowerCAmelCase ) visited.append(_lowerCAmelCase ) _lowercase : List[Any] = s _lowercase : Dict = [] while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: _lowercase : int = s for node in self.graph[s]: if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) _lowercase : str = node[1] break # check if all the children are visited if s == ss: sorted_nodes.append(stack.pop() ) if len(_lowerCAmelCase ) != 0: _lowercase : str = stack[len(_lowerCAmelCase ) - 1] else: _lowercase : int = ss # check if se have reached the starting point if len(_lowerCAmelCase ) == 0: return sorted_nodes def __a ( self ): _lowercase : Tuple = [] _lowercase : Tuple = [] _lowercase : Dict = list(self.graph )[0] stack.append(_lowerCAmelCase ) visited.append(_lowerCAmelCase ) _lowercase : Optional[int] = -2 _lowercase : Tuple = [] _lowercase : Dict = s _lowercase : List[str] = False _lowercase : List[Any] = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: _lowercase : List[Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): _lowercase : Union[str, Any] = len(_lowerCAmelCase ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) _lowercase : Any = node[1] break # check if all the children are visited if s == ss: stack.pop() _lowercase : int = True if len(_lowerCAmelCase ) != 0: _lowercase : List[Any] = stack[len(_lowerCAmelCase ) - 1] else: _lowercase : Tuple = False indirect_parents.append(_lowerCAmelCase ) _lowercase : Dict = s _lowercase : Union[str, Any] = ss # check if se have reached the starting point if len(_lowerCAmelCase ) == 0: return list(_lowerCAmelCase ) def __a ( self ): _lowercase : Optional[Any] = [] _lowercase : Optional[Any] = [] _lowercase : List[Any] = list(self.graph )[0] stack.append(_lowerCAmelCase ) visited.append(_lowerCAmelCase ) _lowercase : Dict = -2 _lowercase : Union[str, Any] = [] _lowercase : int = s _lowercase : List[str] = False _lowercase : Dict = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: _lowercase : List[Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): _lowercase : str = len(_lowerCAmelCase ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) _lowercase : Dict = node[1] break # check if all the children are visited if s == ss: stack.pop() _lowercase : Dict = True if len(_lowerCAmelCase ) != 0: _lowercase : Union[str, Any] = stack[len(_lowerCAmelCase ) - 1] else: _lowercase : Any = False indirect_parents.append(_lowerCAmelCase ) _lowercase : Optional[Any] = s _lowercase : Any = ss # check if se have reached the starting point if len(_lowerCAmelCase ) == 0: return False def __a ( self , _lowerCAmelCase=-2 , _lowerCAmelCase=-1 ): _lowercase : Optional[int] = time() self.dfs(_lowerCAmelCase , _lowerCAmelCase ) _lowercase : Optional[int] = time() return end - begin def __a ( self , _lowerCAmelCase=-2 ): _lowercase : str = time() self.bfs(_lowerCAmelCase ) _lowercase : str = time() return end - begin class lowerCAmelCase_ : def __init__( self ): _lowercase : str = {} def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase=1 ): # check if the u exists if self.graph.get(_lowerCAmelCase ): # if there already is a edge if self.graph[u].count([w, v] ) == 0: self.graph[u].append([w, v] ) else: # if u does not exist _lowercase : Any = [[w, v]] # add the other way if self.graph.get(_lowerCAmelCase ): # if there already is a edge if self.graph[v].count([w, u] ) == 0: self.graph[v].append([w, u] ) else: # if u does not exist _lowercase : List[Any] = [[w, u]] def __a ( self , _lowerCAmelCase , _lowerCAmelCase ): if self.graph.get(_lowerCAmelCase ): for _ in self.graph[u]: if _[1] == v: self.graph[u].remove(_lowerCAmelCase ) # the other way round if self.graph.get(_lowerCAmelCase ): for _ in self.graph[v]: if _[1] == u: self.graph[v].remove(_lowerCAmelCase ) def __a ( self , _lowerCAmelCase=-2 , _lowerCAmelCase=-1 ): if s == d: return [] _lowercase : Dict = [] _lowercase : Dict = [] if s == -2: _lowercase : Tuple = list(self.graph )[0] stack.append(_lowerCAmelCase ) visited.append(_lowerCAmelCase ) _lowercase : str = s while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: _lowercase : Any = s for node in self.graph[s]: if visited.count(node[1] ) < 1: if node[1] == d: visited.append(_lowerCAmelCase ) return visited else: stack.append(node[1] ) visited.append(node[1] ) _lowercase : Union[str, Any] = node[1] break # check if all the children are visited if s == ss: stack.pop() if len(_lowerCAmelCase ) != 0: _lowercase : Any = stack[len(_lowerCAmelCase ) - 1] else: _lowercase : Tuple = ss # check if se have reached the starting point if len(_lowerCAmelCase ) == 0: return visited def __a ( self , _lowerCAmelCase=-1 ): if c == -1: _lowercase : Dict = floor(random() * 1_0_0_0_0 ) + 1_0 for i in range(_lowerCAmelCase ): # every vertex has max 100 edges for _ in range(floor(random() * 1_0_2 ) + 1 ): _lowercase : Optional[Any] = floor(random() * c ) + 1 if n != i: self.add_pair(_lowerCAmelCase , _lowerCAmelCase , 1 ) def __a ( self , _lowerCAmelCase=-2 ): _lowercase : Optional[int] = deque() _lowercase : Any = [] if s == -2: _lowercase : Optional[int] = list(self.graph )[0] d.append(_lowerCAmelCase ) visited.append(_lowerCAmelCase ) while d: _lowercase : Any = d.popleft() if len(self.graph[s] ) != 0: for node in self.graph[s]: if visited.count(node[1] ) < 1: d.append(node[1] ) visited.append(node[1] ) return visited def __a ( self , _lowerCAmelCase ): return len(self.graph[u] ) def __a ( self ): _lowercase : Any = [] _lowercase : Optional[Any] = [] _lowercase : Optional[int] = list(self.graph )[0] stack.append(_lowerCAmelCase ) visited.append(_lowerCAmelCase ) _lowercase : Optional[int] = -2 _lowercase : Dict = [] _lowercase : int = s _lowercase : Dict = False _lowercase : Dict = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: _lowercase : Any = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): _lowercase : str = len(_lowerCAmelCase ) - 1 while len_stack >= 0: if stack[len_stack] == node[1]: anticipating_nodes.add(node[1] ) break else: anticipating_nodes.add(stack[len_stack] ) len_stack -= 1 if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) _lowercase : str = node[1] break # check if all the children are visited if s == ss: stack.pop() _lowercase : Any = True if len(_lowerCAmelCase ) != 0: _lowercase : List[Any] = stack[len(_lowerCAmelCase ) - 1] else: _lowercase : Union[str, Any] = False indirect_parents.append(_lowerCAmelCase ) _lowercase : int = s _lowercase : Any = ss # check if se have reached the starting point if len(_lowerCAmelCase ) == 0: return list(_lowerCAmelCase ) def __a ( self ): _lowercase : Union[str, Any] = [] _lowercase : Dict = [] _lowercase : List[str] = list(self.graph )[0] stack.append(_lowerCAmelCase ) visited.append(_lowerCAmelCase ) _lowercase : Any = -2 _lowercase : Optional[Any] = [] _lowercase : List[str] = s _lowercase : Optional[int] = False _lowercase : Tuple = set() while True: # check if there is any non isolated nodes if len(self.graph[s] ) != 0: _lowercase : Union[str, Any] = s for node in self.graph[s]: if ( visited.count(node[1] ) > 0 and node[1] != parent and indirect_parents.count(node[1] ) > 0 and not on_the_way_back ): _lowercase : Tuple = len(_lowerCAmelCase ) - 1 while len_stack_minus_one >= 0: if stack[len_stack_minus_one] == node[1]: anticipating_nodes.add(node[1] ) break else: return True if visited.count(node[1] ) < 1: stack.append(node[1] ) visited.append(node[1] ) _lowercase : Dict = node[1] break # check if all the children are visited if s == ss: stack.pop() _lowercase : str = True if len(_lowerCAmelCase ) != 0: _lowercase : List[str] = stack[len(_lowerCAmelCase ) - 1] else: _lowercase : Optional[Any] = False indirect_parents.append(_lowerCAmelCase ) _lowercase : Optional[int] = s _lowercase : List[str] = ss # check if se have reached the starting point if len(_lowerCAmelCase ) == 0: return False def __a ( self ): return list(self.graph ) def __a ( self , _lowerCAmelCase=-2 , _lowerCAmelCase=-1 ): _lowercase : Any = time() self.dfs(_lowerCAmelCase , _lowerCAmelCase ) _lowercase : Optional[int] = time() return end - begin def __a ( self , _lowerCAmelCase=-2 ): _lowercase : Dict = time() self.bfs(_lowerCAmelCase ) _lowercase : int = time() return end - begin
66
from __future__ import annotations import math def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> list[int]: if num <= 0: _lowercase : List[str] = F"""{num}: Invalid input, please enter a positive integer.""" raise ValueError(SCREAMING_SNAKE_CASE ) _lowercase : Union[str, Any] = [True] * (num + 1) _lowercase : Union[str, Any] = [] _lowercase : Dict = 2 _lowercase : Union[str, Any] = int(math.sqrt(SCREAMING_SNAKE_CASE ) ) while start <= end: # If start is a prime if sieve[start] is True: prime.append(SCREAMING_SNAKE_CASE ) # Set multiples of start be False for i in range(start * start , num + 1 , SCREAMING_SNAKE_CASE ): if sieve[i] is True: _lowercase : str = False start += 1 for j in range(end + 1 , num + 1 ): if sieve[j] is True: prime.append(SCREAMING_SNAKE_CASE ) return prime if __name__ == "__main__": print(prime_sieve(int(input("Enter a positive integer: ").strip())))
66
1
def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: assert column_title.isupper() _lowercase : Optional[Any] = 0 _lowercase : Optional[Any] = len(SCREAMING_SNAKE_CASE ) - 1 _lowercase : Optional[int] = 0 while index >= 0: _lowercase : Union[str, Any] = (ord(column_title[index] ) - 64) * pow(26 , SCREAMING_SNAKE_CASE ) answer += value power += 1 index -= 1 return answer if __name__ == "__main__": from doctest import testmod testmod()
66
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ConvNextConfig, SegformerImageProcessor, UperNetConfig, UperNetForSemanticSegmentation def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> List[Any]: _lowercase : int = 384 if "tiny" in model_name: _lowercase : Tuple = [3, 3, 9, 3] _lowercase : List[str] = [96, 192, 384, 768] if "small" in model_name: _lowercase : List[str] = [3, 3, 27, 3] _lowercase : Union[str, Any] = [96, 192, 384, 768] if "base" in model_name: _lowercase : List[Any] = [3, 3, 27, 3] _lowercase : Dict = [128, 256, 512, 1_024] _lowercase : Optional[int] = 512 if "large" in model_name: _lowercase : List[str] = [3, 3, 27, 3] _lowercase : List[Any] = [192, 384, 768, 1_536] _lowercase : Tuple = 768 if "xlarge" in model_name: _lowercase : str = [3, 3, 27, 3] _lowercase : List[str] = [256, 512, 1_024, 2_048] _lowercase : Tuple = 1_024 # set label information _lowercase : Dict = 150 _lowercase : Union[str, Any] = 'huggingface/label-files' _lowercase : str = 'ade20k-id2label.json' _lowercase : List[Any] = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , repo_type='dataset' ) , 'r' ) ) _lowercase : Dict = {int(SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} _lowercase : Tuple = {v: k for k, v in idalabel.items()} _lowercase : List[str] = ConvNextConfig( depths=SCREAMING_SNAKE_CASE , hidden_sizes=SCREAMING_SNAKE_CASE , out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) _lowercase : Union[str, Any] = UperNetConfig( backbone_config=SCREAMING_SNAKE_CASE , auxiliary_in_channels=SCREAMING_SNAKE_CASE , num_labels=SCREAMING_SNAKE_CASE , idalabel=SCREAMING_SNAKE_CASE , labelaid=SCREAMING_SNAKE_CASE , ) return config def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : Any = [] # fmt: off # stem rename_keys.append(('backbone.downsample_layers.0.0.weight', 'backbone.embeddings.patch_embeddings.weight') ) rename_keys.append(('backbone.downsample_layers.0.0.bias', 'backbone.embeddings.patch_embeddings.bias') ) rename_keys.append(('backbone.downsample_layers.0.1.weight', 'backbone.embeddings.layernorm.weight') ) rename_keys.append(('backbone.downsample_layers.0.1.bias', 'backbone.embeddings.layernorm.bias') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F"""backbone.stages.{i}.{j}.gamma""", F"""backbone.encoder.stages.{i}.layers.{j}.layer_scale_parameter""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.depthwise_conv.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.dwconv.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.depthwise_conv.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.dwconv.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.norm.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.layernorm.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.norm.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.layernorm.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv1.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv1.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv1.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv1.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv2.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv2.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv2.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv2.bias""") ) if i > 0: rename_keys.append((F"""backbone.downsample_layers.{i}.0.weight""", F"""backbone.encoder.stages.{i}.downsampling_layer.0.weight""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.0.bias""", F"""backbone.encoder.stages.{i}.downsampling_layer.0.bias""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.1.weight""", F"""backbone.encoder.stages.{i}.downsampling_layer.1.weight""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.1.bias""", F"""backbone.encoder.stages.{i}.downsampling_layer.1.bias""") ) rename_keys.append((F"""backbone.norm{i}.weight""", F"""backbone.hidden_states_norms.stage{i+1}.weight""") ) rename_keys.append((F"""backbone.norm{i}.bias""", F"""backbone.hidden_states_norms.stage{i+1}.bias""") ) # decode head rename_keys.extend( [ ('decode_head.conv_seg.weight', 'decode_head.classifier.weight'), ('decode_head.conv_seg.bias', 'decode_head.classifier.bias'), ('auxiliary_head.conv_seg.weight', 'auxiliary_head.classifier.weight'), ('auxiliary_head.conv_seg.bias', 'auxiliary_head.classifier.bias'), ] ) # fmt: on return rename_keys def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Optional[int]: _lowercase : Any = dct.pop(SCREAMING_SNAKE_CASE ) _lowercase : Any = val def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Any: _lowercase : List[Any] = { 'upernet-convnext-tiny': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_tiny_fp16_512x512_160k_ade20k/upernet_convnext_tiny_fp16_512x512_160k_ade20k_20220227_124553-cad485de.pth', 'upernet-convnext-small': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_small_fp16_512x512_160k_ade20k/upernet_convnext_small_fp16_512x512_160k_ade20k_20220227_131208-1b1e394f.pth', 'upernet-convnext-base': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_base_fp16_512x512_160k_ade20k/upernet_convnext_base_fp16_512x512_160k_ade20k_20220227_181227-02a24fc6.pth', 'upernet-convnext-large': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_large_fp16_640x640_160k_ade20k/upernet_convnext_large_fp16_640x640_160k_ade20k_20220226_040532-e57aa54d.pth', 'upernet-convnext-xlarge': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_xlarge_fp16_640x640_160k_ade20k/upernet_convnext_xlarge_fp16_640x640_160k_ade20k_20220226_080344-95fc38c2.pth', } _lowercase : Optional[int] = model_name_to_url[model_name] _lowercase : str = torch.hub.load_state_dict_from_url(SCREAMING_SNAKE_CASE , map_location='cpu' )['state_dict'] _lowercase : Optional[int] = get_upernet_config(SCREAMING_SNAKE_CASE ) _lowercase : Tuple = UperNetForSemanticSegmentation(SCREAMING_SNAKE_CASE ) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): _lowercase : List[Any] = state_dict.pop(SCREAMING_SNAKE_CASE ) if "bn" in key: _lowercase : Any = key.replace('bn' , 'batch_norm' ) _lowercase : Any = val # rename keys _lowercase : int = create_rename_keys(SCREAMING_SNAKE_CASE ) for src, dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) model.load_state_dict(SCREAMING_SNAKE_CASE ) # verify on image _lowercase : Union[str, Any] = 'https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg' _lowercase : Any = Image.open(requests.get(SCREAMING_SNAKE_CASE , stream=SCREAMING_SNAKE_CASE ).raw ).convert('RGB' ) _lowercase : Tuple = SegformerImageProcessor() _lowercase : Tuple = processor(SCREAMING_SNAKE_CASE , return_tensors='pt' ).pixel_values with torch.no_grad(): _lowercase : Dict = model(SCREAMING_SNAKE_CASE ) if model_name == "upernet-convnext-tiny": _lowercase : Dict = torch.tensor( [[-8.8110, -8.8110, -8.6521], [-8.8110, -8.8110, -8.6521], [-8.7746, -8.7746, -8.6130]] ) elif model_name == "upernet-convnext-small": _lowercase : Union[str, Any] = torch.tensor( [[-8.8236, -8.8236, -8.6771], [-8.8236, -8.8236, -8.6771], [-8.7638, -8.7638, -8.6240]] ) elif model_name == "upernet-convnext-base": _lowercase : Dict = torch.tensor( [[-8.8558, -8.8558, -8.6905], [-8.8558, -8.8558, -8.6905], [-8.7669, -8.7669, -8.6021]] ) elif model_name == "upernet-convnext-large": _lowercase : Optional[int] = torch.tensor( [[-8.6660, -8.6660, -8.6210], [-8.6660, -8.6660, -8.6210], [-8.6310, -8.6310, -8.5964]] ) elif model_name == "upernet-convnext-xlarge": _lowercase : str = torch.tensor( [[-8.4980, -8.4980, -8.3977], [-8.4980, -8.4980, -8.3977], [-8.4379, -8.4379, -8.3412]] ) print('Logits:' , outputs.logits[0, 0, :3, :3] ) assert torch.allclose(outputs.logits[0, 0, :3, :3] , SCREAMING_SNAKE_CASE , atol=1E-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(SCREAMING_SNAKE_CASE ) print(F"""Saving processor to {pytorch_dump_folder_path}""" ) processor.save_pretrained(SCREAMING_SNAKE_CASE ) if push_to_hub: print(F"""Pushing model and processor for {model_name} to hub""" ) model.push_to_hub(F"""openmmlab/{model_name}""" ) processor.push_to_hub(F"""openmmlab/{model_name}""" ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="upernet-convnext-tiny", type=str, choices=[f'''upernet-convnext-{size}''' for size in ["tiny", "small", "base", "large", "xlarge"]], help="Name of the ConvNext UperNet model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) UpperCamelCase = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
66
1
import logging import os import threading import time try: import warnings except ImportError: UpperCamelCase = None try: import msvcrt except ImportError: UpperCamelCase = None try: import fcntl except ImportError: UpperCamelCase = None # Backward compatibility # ------------------------------------------------ try: TimeoutError except NameError: UpperCamelCase = OSError # Data # ------------------------------------------------ UpperCamelCase = [ "Timeout", "BaseFileLock", "WindowsFileLock", "UnixFileLock", "SoftFileLock", "FileLock", ] UpperCamelCase = "3.0.12" UpperCamelCase = None def __magic_name__ ( ) -> List[Any]: global _logger _lowercase : Optional[int] = _logger or logging.getLogger(__name__ ) return _logger class lowerCAmelCase_ ( __snake_case ): def __init__( self , _lowerCAmelCase ): _lowercase : List[Any] = lock_file return None def __str__( self ): _lowercase : Tuple = F"""The file lock '{self.lock_file}' could not be acquired.""" return temp class lowerCAmelCase_ : def __init__( self , _lowerCAmelCase ): _lowercase : Any = lock return None def __enter__( self ): return self.lock def __exit__( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): self.lock.release() return None class lowerCAmelCase_ : def __init__( self , _lowerCAmelCase , _lowerCAmelCase=-1 , _lowerCAmelCase=None ): _lowercase : Tuple = max_filename_length if max_filename_length is not None else 2_5_5 # Hash the filename if it's too long _lowercase : Union[str, Any] = self.hash_filename_if_too_long(_lowerCAmelCase , _lowerCAmelCase ) # The path to the lock file. _lowercase : int = lock_file # The file descriptor for the *_lock_file* as it is returned by the # os.open() function. # This file lock is only NOT None, if the object currently holds the # lock. _lowercase : List[Any] = None # The default timeout value. _lowercase : Union[str, Any] = timeout # We use this lock primarily for the lock counter. _lowercase : Any = threading.Lock() # The lock counter is used for implementing the nested locking # mechanism. Whenever the lock is acquired, the counter is increased and # the lock is only released, when this value is 0 again. _lowercase : Optional[Any] = 0 return None @property def __a ( self ): return self._lock_file @property def __a ( self ): return self._timeout @timeout.setter def __a ( self , _lowerCAmelCase ): _lowercase : Tuple = float(_lowerCAmelCase ) return None def __a ( self ): raise NotImplementedError() def __a ( self ): raise NotImplementedError() @property def __a ( self ): return self._lock_file_fd is not None def __a ( self , _lowerCAmelCase=None , _lowerCAmelCase=0.05 ): # Use the default timeout, if no timeout is provided. if timeout is None: _lowercase : Any = self.timeout # Increment the number right at the beginning. # We can still undo it, if something fails. with self._thread_lock: self._lock_counter += 1 _lowercase : Tuple = id(self ) _lowercase : Union[str, Any] = self._lock_file _lowercase : Tuple = time.time() try: while True: with self._thread_lock: if not self.is_locked: logger().debug(F"""Attempting to acquire lock {lock_id} on {lock_filename}""" ) self._acquire() if self.is_locked: logger().debug(F"""Lock {lock_id} acquired on {lock_filename}""" ) break elif timeout >= 0 and time.time() - start_time > timeout: logger().debug(F"""Timeout on acquiring lock {lock_id} on {lock_filename}""" ) raise Timeout(self._lock_file ) else: logger().debug( F"""Lock {lock_id} not acquired on {lock_filename}, waiting {poll_intervall} seconds ...""" ) time.sleep(_lowerCAmelCase ) except: # noqa # Something did go wrong, so decrement the counter. with self._thread_lock: _lowercase : Dict = max(0 , self._lock_counter - 1 ) raise return _Acquire_ReturnProxy(lock=self ) def __a ( self , _lowerCAmelCase=False ): with self._thread_lock: if self.is_locked: self._lock_counter -= 1 if self._lock_counter == 0 or force: _lowercase : Tuple = id(self ) _lowercase : Optional[Any] = self._lock_file logger().debug(F"""Attempting to release lock {lock_id} on {lock_filename}""" ) self._release() _lowercase : Union[str, Any] = 0 logger().debug(F"""Lock {lock_id} released on {lock_filename}""" ) return None def __enter__( self ): self.acquire() return self def __exit__( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): self.release() return None def __del__( self ): self.release(force=_lowerCAmelCase ) return None def __a ( self , _lowerCAmelCase , _lowerCAmelCase ): _lowercase : Optional[Any] = os.path.basename(_lowerCAmelCase ) if len(_lowerCAmelCase ) > max_length and max_length > 0: _lowercase : Optional[int] = os.path.dirname(_lowerCAmelCase ) _lowercase : int = str(hash(_lowerCAmelCase ) ) _lowercase : Union[str, Any] = filename[: max_length - len(_lowerCAmelCase ) - 8] + '...' + hashed_filename + '.lock' return os.path.join(_lowerCAmelCase , _lowerCAmelCase ) else: return path class lowerCAmelCase_ ( __snake_case ): def __init__( self , _lowerCAmelCase , _lowerCAmelCase=-1 , _lowerCAmelCase=None ): from .file_utils import relative_to_absolute_path super().__init__(_lowerCAmelCase , timeout=_lowerCAmelCase , max_filename_length=_lowerCAmelCase ) _lowercase : Optional[Any] = '\\\\?\\' + relative_to_absolute_path(self.lock_file ) def __a ( self ): _lowercase : str = os.O_RDWR | os.O_CREAT | os.O_TRUNC try: _lowercase : Dict = os.open(self._lock_file , _lowerCAmelCase ) except OSError: pass else: try: msvcrt.locking(_lowerCAmelCase , msvcrt.LK_NBLCK , 1 ) except OSError: os.close(_lowerCAmelCase ) else: _lowercase : List[str] = fd return None def __a ( self ): _lowercase : Tuple = self._lock_file_fd _lowercase : Union[str, Any] = None msvcrt.locking(_lowerCAmelCase , msvcrt.LK_UNLCK , 1 ) os.close(_lowerCAmelCase ) try: os.remove(self._lock_file ) # Probably another instance of the application # that acquired the file lock. except OSError: pass return None class lowerCAmelCase_ ( __snake_case ): def __init__( self , _lowerCAmelCase , _lowerCAmelCase=-1 , _lowerCAmelCase=None ): _lowercase : List[str] = os.statvfs(os.path.dirname(_lowerCAmelCase ) ).f_namemax super().__init__(_lowerCAmelCase , timeout=_lowerCAmelCase , max_filename_length=_lowerCAmelCase ) def __a ( self ): _lowercase : List[str] = os.O_RDWR | os.O_CREAT | os.O_TRUNC _lowercase : List[str] = os.open(self._lock_file , _lowerCAmelCase ) try: fcntl.flock(_lowerCAmelCase , fcntl.LOCK_EX | fcntl.LOCK_NB ) except OSError: os.close(_lowerCAmelCase ) else: _lowercase : Optional[Any] = fd return None def __a ( self ): # Do not remove the lockfile: # # https://github.com/benediktschmitt/py-filelock/issues/31 # https://stackoverflow.com/questions/17708885/flock-removing-locked-file-without-race-condition _lowercase : Tuple = self._lock_file_fd _lowercase : Optional[Any] = None fcntl.flock(_lowerCAmelCase , fcntl.LOCK_UN ) os.close(_lowerCAmelCase ) return None class lowerCAmelCase_ ( __snake_case ): def __a ( self ): _lowercase : str = os.O_WRONLY | os.O_CREAT | os.O_EXCL | os.O_TRUNC try: _lowercase : int = os.open(self._lock_file , _lowerCAmelCase ) except OSError: pass else: _lowercase : str = fd return None def __a ( self ): os.close(self._lock_file_fd ) _lowercase : Union[str, Any] = None try: os.remove(self._lock_file ) # The file is already deleted and that's what we want. except OSError: pass return None UpperCamelCase = None if msvcrt: UpperCamelCase = WindowsFileLock elif fcntl: UpperCamelCase = UnixFileLock else: UpperCamelCase = SoftFileLock if warnings is not None: warnings.warn("only soft file lock is available")
66
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto.configuration_auto import CONFIG_MAPPING UpperCamelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : Any = "upernet" def __init__( self , _lowerCAmelCase=None , _lowerCAmelCase=5_1_2 , _lowerCAmelCase=0.02 , _lowerCAmelCase=[1, 2, 3, 6] , _lowerCAmelCase=True , _lowerCAmelCase=0.4 , _lowerCAmelCase=3_8_4 , _lowerCAmelCase=2_5_6 , _lowerCAmelCase=1 , _lowerCAmelCase=False , _lowerCAmelCase=2_5_5 , **_lowerCAmelCase , ): super().__init__(**_lowerCAmelCase ) if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.' ) _lowercase : Optional[Any] = CONFIG_MAPPING['resnet'](out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): _lowercase : List[Any] = backbone_config.get('model_type' ) _lowercase : str = CONFIG_MAPPING[backbone_model_type] _lowercase : Tuple = config_class.from_dict(_lowerCAmelCase ) _lowercase : Optional[Any] = backbone_config _lowercase : Any = hidden_size _lowercase : Any = initializer_range _lowercase : Tuple = pool_scales _lowercase : List[Any] = use_auxiliary_head _lowercase : Optional[Any] = auxiliary_loss_weight _lowercase : Any = auxiliary_in_channels _lowercase : Any = auxiliary_channels _lowercase : List[str] = auxiliary_num_convs _lowercase : List[str] = auxiliary_concat_input _lowercase : Tuple = loss_ignore_index def __a ( self ): _lowercase : str = copy.deepcopy(self.__dict__ ) _lowercase : Tuple = self.backbone_config.to_dict() _lowercase : int = self.__class__.model_type return output
66
1
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL UpperCamelCase = logging.get_logger(__name__) def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> List[List[ImageInput]]: if isinstance(SCREAMING_SNAKE_CASE , (list, tuple) ) and isinstance(videos[0] , (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(SCREAMING_SNAKE_CASE , (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(SCREAMING_SNAKE_CASE ): return [[videos]] raise ValueError(F"""Could not make batched video from {videos}""" ) class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : List[Any] = ["pixel_values"] def __init__( self , _lowerCAmelCase = True , _lowerCAmelCase = None , _lowerCAmelCase = PILImageResampling.BILINEAR , _lowerCAmelCase = True , _lowerCAmelCase = None , _lowerCAmelCase = True , _lowerCAmelCase = 1 / 2_5_5 , _lowerCAmelCase = True , _lowerCAmelCase = None , _lowerCAmelCase = None , **_lowerCAmelCase , ): super().__init__(**_lowerCAmelCase ) _lowercase : List[str] = size if size is not None else {'shortest_edge': 2_2_4} _lowercase : Union[str, Any] = get_size_dict(_lowerCAmelCase , default_to_square=_lowerCAmelCase ) _lowercase : Optional[int] = crop_size if crop_size is not None else {'height': 2_2_4, 'width': 2_2_4} _lowercase : Optional[Any] = get_size_dict(_lowerCAmelCase , param_name='crop_size' ) _lowercase : str = do_resize _lowercase : Optional[int] = size _lowercase : Optional[Any] = do_center_crop _lowercase : Dict = crop_size _lowercase : Dict = resample _lowercase : List[Any] = do_rescale _lowercase : List[Any] = rescale_factor _lowercase : Optional[Any] = do_normalize _lowercase : Optional[Any] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _lowercase : Dict = image_std if image_std is not None else IMAGENET_STANDARD_STD def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = PILImageResampling.BILINEAR , _lowerCAmelCase = None , **_lowerCAmelCase , ): _lowercase : Optional[Any] = get_size_dict(_lowerCAmelCase , default_to_square=_lowerCAmelCase ) if "shortest_edge" in size: _lowercase : str = get_resize_output_image_size(_lowerCAmelCase , size['shortest_edge'] , default_to_square=_lowerCAmelCase ) elif "height" in size and "width" in size: _lowercase : Tuple = (size['height'], size['width']) else: raise ValueError(F"""Size must have 'height' and 'width' or 'shortest_edge' as keys. Got {size.keys()}""" ) return resize(_lowerCAmelCase , size=_lowerCAmelCase , resample=_lowerCAmelCase , data_format=_lowerCAmelCase , **_lowerCAmelCase ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = None , **_lowerCAmelCase , ): _lowercase : Optional[int] = get_size_dict(_lowerCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(F"""Size must have 'height' and 'width' as keys. Got {size.keys()}""" ) return center_crop(_lowerCAmelCase , size=(size['height'], size['width']) , data_format=_lowerCAmelCase , **_lowerCAmelCase ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = None , **_lowerCAmelCase , ): return rescale(_lowerCAmelCase , scale=_lowerCAmelCase , data_format=_lowerCAmelCase , **_lowerCAmelCase ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = None , **_lowerCAmelCase , ): return normalize(_lowerCAmelCase , mean=_lowerCAmelCase , std=_lowerCAmelCase , data_format=_lowerCAmelCase , **_lowerCAmelCase ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = ChannelDimension.FIRST , ): if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. _lowercase : str = to_numpy_array(_lowerCAmelCase ) if do_resize: _lowercase : List[Any] = self.resize(image=_lowerCAmelCase , size=_lowerCAmelCase , resample=_lowerCAmelCase ) if do_center_crop: _lowercase : Optional[int] = self.center_crop(_lowerCAmelCase , size=_lowerCAmelCase ) if do_rescale: _lowercase : int = self.rescale(image=_lowerCAmelCase , scale=_lowerCAmelCase ) if do_normalize: _lowercase : Any = self.normalize(image=_lowerCAmelCase , mean=_lowerCAmelCase , std=_lowerCAmelCase ) _lowercase : Union[str, Any] = to_channel_dimension_format(_lowerCAmelCase , _lowerCAmelCase ) return image def __a ( self , _lowerCAmelCase , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = ChannelDimension.FIRST , **_lowerCAmelCase , ): _lowercase : Dict = do_resize if do_resize is not None else self.do_resize _lowercase : Optional[Any] = resample if resample is not None else self.resample _lowercase : List[str] = do_center_crop if do_center_crop is not None else self.do_center_crop _lowercase : List[Any] = do_rescale if do_rescale is not None else self.do_rescale _lowercase : Tuple = rescale_factor if rescale_factor is not None else self.rescale_factor _lowercase : Any = do_normalize if do_normalize is not None else self.do_normalize _lowercase : Dict = image_mean if image_mean is not None else self.image_mean _lowercase : List[str] = image_std if image_std is not None else self.image_std _lowercase : Optional[Any] = size if size is not None else self.size _lowercase : Union[str, Any] = get_size_dict(_lowerCAmelCase , default_to_square=_lowerCAmelCase ) _lowercase : int = crop_size if crop_size is not None else self.crop_size _lowercase : List[Any] = get_size_dict(_lowerCAmelCase , param_name='crop_size' ) if not valid_images(_lowerCAmelCase ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) _lowercase : Optional[Any] = make_batched(_lowerCAmelCase ) _lowercase : Tuple = [ [ self._preprocess_image( image=_lowerCAmelCase , do_resize=_lowerCAmelCase , size=_lowerCAmelCase , resample=_lowerCAmelCase , do_center_crop=_lowerCAmelCase , crop_size=_lowerCAmelCase , do_rescale=_lowerCAmelCase , rescale_factor=_lowerCAmelCase , do_normalize=_lowerCAmelCase , image_mean=_lowerCAmelCase , image_std=_lowerCAmelCase , data_format=_lowerCAmelCase , ) for img in video ] for video in videos ] _lowercase : Any = {'pixel_values': videos} return BatchFeature(data=_lowerCAmelCase , tensor_type=_lowerCAmelCase )
66
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: if a < 0 or b < 0: raise ValueError('the value of both inputs must be positive' ) _lowercase : str = str(bin(SCREAMING_SNAKE_CASE ) )[2:] # remove the leading "0b" _lowercase : int = str(bin(SCREAMING_SNAKE_CASE ) )[2:] # remove the leading "0b" _lowercase : List[Any] = max(len(SCREAMING_SNAKE_CASE ) , len(SCREAMING_SNAKE_CASE ) ) return "0b" + "".join( str(int(char_a == '1' and char_b == '1' ) ) for char_a, char_b in zip(a_binary.zfill(SCREAMING_SNAKE_CASE ) , b_binary.zfill(SCREAMING_SNAKE_CASE ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
66
1
# Copyright 2022 The HuggingFace Team and The OpenBMB Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCamelCase = { "configuration_cpmant": ["CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP", "CpmAntConfig"], "tokenization_cpmant": ["CpmAntTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST", "CpmAntForCausalLM", "CpmAntModel", "CpmAntPreTrainedModel", ] if TYPE_CHECKING: from .configuration_cpmant import CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP, CpmAntConfig from .tokenization_cpmant import CpmAntTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_cpmant import ( CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST, CpmAntForCausalLM, CpmAntModel, CpmAntPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class lowerCAmelCase_ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCamelCase : int = IFInpaintingSuperResolutionPipeline _UpperCamelCase : Any = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"width", "height"} _UpperCamelCase : Union[str, Any] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"original_image"} ) _UpperCamelCase : Tuple = PipelineTesterMixin.required_optional_params - {"latents"} def __a ( self ): return self._get_superresolution_dummy_components() def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : int = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Union[str, Any] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : Union[str, Any] = floats_tensor((1, 3, 1_6, 1_6) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : Optional[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'original_image': original_image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def __a ( self ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def __a ( self ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != 'cuda' , reason='float16 requires CUDA' ) def __a ( self ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __a ( self ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __a ( self ): self._test_save_load_local() def __a ( self ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
66
1
UpperCamelCase = frozenset( [ "prompt", "height", "width", "guidance_scale", "negative_prompt", "prompt_embeds", "negative_prompt_embeds", "cross_attention_kwargs", ] ) UpperCamelCase = frozenset(["prompt", "negative_prompt"]) UpperCamelCase = frozenset([]) UpperCamelCase = frozenset(["image"]) UpperCamelCase = frozenset( [ "image", "height", "width", "guidance_scale", ] ) UpperCamelCase = frozenset(["image"]) UpperCamelCase = frozenset( [ "prompt", "image", "height", "width", "guidance_scale", "negative_prompt", "prompt_embeds", "negative_prompt_embeds", ] ) UpperCamelCase = frozenset(["prompt", "image", "negative_prompt"]) UpperCamelCase = frozenset( [ # Text guided image variation with an image mask "prompt", "image", "mask_image", "height", "width", "guidance_scale", "negative_prompt", "prompt_embeds", "negative_prompt_embeds", ] ) UpperCamelCase = frozenset(["prompt", "image", "mask_image", "negative_prompt"]) UpperCamelCase = frozenset( [ # image variation with an image mask "image", "mask_image", "height", "width", "guidance_scale", ] ) UpperCamelCase = frozenset(["image", "mask_image"]) UpperCamelCase = frozenset( [ "example_image", "image", "mask_image", "height", "width", "guidance_scale", ] ) UpperCamelCase = frozenset(["example_image", "image", "mask_image"]) UpperCamelCase = frozenset(["class_labels"]) UpperCamelCase = frozenset(["class_labels"]) UpperCamelCase = frozenset(["batch_size"]) UpperCamelCase = frozenset([]) UpperCamelCase = frozenset(["batch_size"]) UpperCamelCase = frozenset([]) UpperCamelCase = frozenset( [ "prompt", "audio_length_in_s", "guidance_scale", "negative_prompt", "prompt_embeds", "negative_prompt_embeds", "cross_attention_kwargs", ] ) UpperCamelCase = frozenset(["prompt", "negative_prompt"]) UpperCamelCase = frozenset(["input_tokens"]) UpperCamelCase = frozenset(["input_tokens"])
66
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> tuple[float, float]: # Check if the input is valid if not len(SCREAMING_SNAKE_CASE ) == len(SCREAMING_SNAKE_CASE ) == 3: raise ValueError('Please enter a valid equation.' ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError('Both a & b of two equations can\'t be zero.' ) # Extract the coefficients _lowercase , _lowercase , _lowercase : Tuple = equationa _lowercase , _lowercase , _lowercase : Dict = equationa # Calculate the determinants of the matrices _lowercase : str = aa * ba - aa * ba _lowercase : Any = ca * ba - ca * ba _lowercase : Optional[int] = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError('Infinite solutions. (Consistent system)' ) else: raise ValueError('No solution. (Inconsistent system)' ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: _lowercase : Union[str, Any] = determinant_x / determinant _lowercase : Tuple = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
66
1
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : str = ShapEPipeline _UpperCamelCase : Any = ["prompt"] _UpperCamelCase : int = ["prompt"] _UpperCamelCase : Union[str, Any] = [ "num_images_per_prompt", "num_inference_steps", "generator", "latents", "guidance_scale", "frame_size", "output_type", "return_dict", ] _UpperCamelCase : Optional[Any] = False @property def __a ( self ): return 3_2 @property def __a ( self ): return 3_2 @property def __a ( self ): return self.time_input_dim * 4 @property def __a ( self ): return 8 @property def __a ( self ): _lowercase : Any = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) return tokenizer @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) return CLIPTextModelWithProjection(_lowerCAmelCase ) @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = { 'num_attention_heads': 2, 'attention_head_dim': 1_6, 'embedding_dim': self.time_input_dim, 'num_embeddings': 3_2, 'embedding_proj_dim': self.text_embedder_hidden_size, 'time_embed_dim': self.time_embed_dim, 'num_layers': 1, 'clip_embed_dim': self.time_input_dim * 2, 'additional_embeddings': 0, 'time_embed_act_fn': 'gelu', 'norm_in_type': 'layer', 'encoder_hid_proj_type': None, 'added_emb_type': None, } _lowercase : Optional[Any] = PriorTransformer(**_lowerCAmelCase ) return model @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = { 'param_shapes': ( (self.renderer_dim, 9_3), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), 'd_latent': self.time_input_dim, 'd_hidden': self.renderer_dim, 'n_output': 1_2, 'background': ( 0.1, 0.1, 0.1, ), } _lowercase : List[Any] = ShapERenderer(**_lowerCAmelCase ) return model def __a ( self ): _lowercase : Optional[Any] = self.dummy_prior _lowercase : Dict = self.dummy_text_encoder _lowercase : List[str] = self.dummy_tokenizer _lowercase : Union[str, Any] = self.dummy_renderer _lowercase : List[str] = HeunDiscreteScheduler( beta_schedule='exp' , num_train_timesteps=1_0_2_4 , prediction_type='sample' , use_karras_sigmas=_lowerCAmelCase , clip_sample=_lowerCAmelCase , clip_sample_range=1.0 , ) _lowercase : List[str] = { 'prior': prior, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'renderer': renderer, 'scheduler': scheduler, } return components def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : Optional[Any] = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Optional[int] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : List[Any] = { 'prompt': 'horse', 'generator': generator, 'num_inference_steps': 1, 'frame_size': 3_2, 'output_type': 'np', } return inputs def __a ( self ): _lowercase : Optional[int] = 'cpu' _lowercase : List[Any] = self.get_dummy_components() _lowercase : Tuple = self.pipeline_class(**_lowerCAmelCase ) _lowercase : Union[str, Any] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Union[str, Any] = pipe(**self.get_dummy_inputs(_lowerCAmelCase ) ) _lowercase : str = output.images[0] _lowercase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (2_0, 3_2, 3_2, 3) _lowercase : str = np.array( [ 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __a ( self ): # NOTE: Larger batch sizes cause this test to timeout, only test on smaller batches self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def __a ( self ): _lowercase : List[Any] = torch_device == 'cpu' _lowercase : Any = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=_lowerCAmelCase , relax_max_difference=_lowerCAmelCase , ) def __a ( self ): _lowercase : Union[str, Any] = self.get_dummy_components() _lowercase : Optional[int] = self.pipeline_class(**_lowerCAmelCase ) _lowercase : Any = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : str = 1 _lowercase : Optional[int] = 2 _lowercase : List[str] = self.get_dummy_inputs(_lowerCAmelCase ) for key in inputs.keys(): if key in self.batch_params: _lowercase : int = batch_size * [inputs[key]] _lowercase : Optional[int] = pipe(**_lowerCAmelCase , num_images_per_prompt=_lowerCAmelCase )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ): _lowercase : List[str] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/shap_e/test_shap_e_np_out.npy' ) _lowercase : Any = ShapEPipeline.from_pretrained('openai/shap-e' ) _lowercase : List[str] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Tuple = torch.Generator(device=_lowerCAmelCase ).manual_seed(0 ) _lowercase : int = pipe( 'a shark' , generator=_lowerCAmelCase , guidance_scale=15.0 , num_inference_steps=6_4 , frame_size=6_4 , output_type='np' , ).images[0] assert images.shape == (2_0, 6_4, 6_4, 3) assert_mean_pixel_difference(_lowerCAmelCase , _lowerCAmelCase )
66
def __magic_name__ ( SCREAMING_SNAKE_CASE = 50 ) -> int: _lowercase : Optional[int] = [[0] * 3 for _ in range(length + 1 )] for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): different_colour_ways_number[row_length][tile_length - 2] += ( different_colour_ways_number[row_length - tile_start - tile_length][ tile_length - 2 ] + 1 ) return sum(different_colour_ways_number[length] ) if __name__ == "__main__": print(f'''{solution() = }''')
66
1
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaImgaImgPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : Union[str, Any] = KandinskyVaaImgaImgPipeline _UpperCamelCase : Optional[int] = ["image_embeds", "negative_image_embeds", "image"] _UpperCamelCase : Any = [ "image_embeds", "negative_image_embeds", "image", ] _UpperCamelCase : int = [ "generator", "height", "width", "strength", "guidance_scale", "num_inference_steps", "return_dict", "guidance_scale", "num_images_per_prompt", "output_type", "return_dict", ] _UpperCamelCase : str = False @property def __a ( self ): return 3_2 @property def __a ( self ): return 3_2 @property def __a ( self ): return self.time_input_dim @property def __a ( self ): return self.time_input_dim * 4 @property def __a ( self ): return 1_0_0 @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Tuple = { 'in_channels': 4, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'image', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } _lowercase : str = UNetaDConditionModel(**_lowerCAmelCase ) return model @property def __a ( self ): return { "block_out_channels": [3_2, 6_4], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 1_2, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = VQModel(**self.dummy_movq_kwargs ) return model def __a ( self ): _lowercase : Union[str, Any] = self.dummy_unet _lowercase : List[Any] = self.dummy_movq _lowercase : List[Any] = { 'num_train_timesteps': 1_0_0_0, 'beta_schedule': 'linear', 'beta_start': 0.0_00_85, 'beta_end': 0.0_12, 'clip_sample': False, 'set_alpha_to_one': False, 'steps_offset': 0, 'prediction_type': 'epsilon', 'thresholding': False, } _lowercase : str = DDIMScheduler(**_lowerCAmelCase ) _lowercase : Union[str, Any] = { 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): _lowercase : Optional[int] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : int = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( _lowerCAmelCase ) # create init_image _lowercase : Optional[int] = floats_tensor((1, 3, 6_4, 6_4) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : Dict = image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowercase : List[Any] = Image.fromarray(np.uinta(_lowerCAmelCase ) ).convert('RGB' ).resize((2_5_6, 2_5_6) ) if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : List[str] = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Optional[int] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : List[str] = { 'image': init_image, 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'generator': generator, 'height': 6_4, 'width': 6_4, 'num_inference_steps': 1_0, 'guidance_scale': 7.0, 'strength': 0.2, 'output_type': 'np', } return inputs def __a ( self ): _lowercase : Optional[int] = 'cpu' _lowercase : Dict = self.get_dummy_components() _lowercase : List[str] = self.pipeline_class(**_lowerCAmelCase ) _lowercase : Optional[int] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Dict = pipe(**self.get_dummy_inputs(_lowerCAmelCase ) ) _lowercase : int = output.images _lowercase : List[Any] = pipe( **self.get_dummy_inputs(_lowerCAmelCase ) , return_dict=_lowerCAmelCase , )[0] _lowercase : List[str] = image[0, -3:, -3:, -1] _lowercase : Dict = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) _lowercase : Any = np.array( [0.6_19_97_78, 0.63_98_44_06, 0.46_14_57_85, 0.62_94_49_84, 0.5_62_22_15, 0.47_30_61_32, 0.47_44_14_56, 0.4_60_76_06, 0.48_71_92_63] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), F""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), F""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" @slow @require_torch_gpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ): _lowercase : Optional[int] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinskyv22/kandinskyv22_img2img_frog.npy' ) _lowercase : List[Any] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/cat.png' ) _lowercase : List[Any] = 'A red cartoon frog, 4k' _lowercase : Optional[int] = KandinskyVaaPriorPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-prior' , torch_dtype=torch.floataa ) pipe_prior.to(_lowerCAmelCase ) _lowercase : str = KandinskyVaaImgaImgPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-decoder' , torch_dtype=torch.floataa ) _lowercase : Dict = pipeline.to(_lowerCAmelCase ) pipeline.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : str = torch.Generator(device='cpu' ).manual_seed(0 ) _lowercase , _lowercase : Optional[int] = pipe_prior( _lowerCAmelCase , generator=_lowerCAmelCase , num_inference_steps=5 , negative_prompt='' , ).to_tuple() _lowercase : int = pipeline( image=_lowerCAmelCase , image_embeds=_lowerCAmelCase , negative_image_embeds=_lowerCAmelCase , generator=_lowerCAmelCase , num_inference_steps=1_0_0 , height=7_6_8 , width=7_6_8 , strength=0.2 , output_type='np' , ) _lowercase : str = output.images[0] assert image.shape == (7_6_8, 7_6_8, 3) assert_mean_pixel_difference(_lowerCAmelCase , _lowerCAmelCase )
66
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) UpperCamelCase = { "configuration_convnext": ["CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP", "ConvNextConfig", "ConvNextOnnxConfig"] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["ConvNextFeatureExtractor"] UpperCamelCase = ["ConvNextImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST", "ConvNextForImageClassification", "ConvNextModel", "ConvNextPreTrainedModel", "ConvNextBackbone", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "TFConvNextForImageClassification", "TFConvNextModel", "TFConvNextPreTrainedModel", ] if TYPE_CHECKING: from .configuration_convnext import CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvNextConfig, ConvNextOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_convnext import ConvNextFeatureExtractor from .image_processing_convnext import ConvNextImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convnext import ( CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvNextBackbone, ConvNextForImageClassification, ConvNextModel, ConvNextPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convnext import TFConvNextForImageClassification, TFConvNextModel, TFConvNextPreTrainedModel else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure)
66
1
from __future__ import annotations import math def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> list[int]: if num <= 0: _lowercase : List[str] = F"""{num}: Invalid input, please enter a positive integer.""" raise ValueError(SCREAMING_SNAKE_CASE ) _lowercase : Union[str, Any] = [True] * (num + 1) _lowercase : Union[str, Any] = [] _lowercase : Dict = 2 _lowercase : Union[str, Any] = int(math.sqrt(SCREAMING_SNAKE_CASE ) ) while start <= end: # If start is a prime if sieve[start] is True: prime.append(SCREAMING_SNAKE_CASE ) # Set multiples of start be False for i in range(start * start , num + 1 , SCREAMING_SNAKE_CASE ): if sieve[i] is True: _lowercase : str = False start += 1 for j in range(end + 1 , num + 1 ): if sieve[j] is True: prime.append(SCREAMING_SNAKE_CASE ) return prime if __name__ == "__main__": print(prime_sieve(int(input("Enter a positive integer: ").strip())))
66
def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> bool: if p < 2: raise ValueError('p should not be less than 2!' ) elif p == 2: return True _lowercase : Optional[Any] = 4 _lowercase : Tuple = (1 << p) - 1 for _ in range(p - 2 ): _lowercase : Union[str, Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
66
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = { "configuration_poolformer": [ "POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "PoolFormerConfig", "PoolFormerOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["PoolFormerFeatureExtractor"] UpperCamelCase = ["PoolFormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "PoolFormerForImageClassification", "PoolFormerModel", "PoolFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure)
66
import random def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = False ) -> dict: _lowercase : dict = {i: [] for i in range(SCREAMING_SNAKE_CASE )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(SCREAMING_SNAKE_CASE ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(SCREAMING_SNAKE_CASE ): for j in range(i + 1 , SCREAMING_SNAKE_CASE ): if random.random() < probability: graph[i].append(SCREAMING_SNAKE_CASE ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(SCREAMING_SNAKE_CASE ) return graph def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> dict: return { i: [j for j in range(SCREAMING_SNAKE_CASE ) if i != j] for i in range(SCREAMING_SNAKE_CASE ) } if __name__ == "__main__": import doctest doctest.testmod()
66
1
def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> bool: if p < 2: raise ValueError('p should not be less than 2!' ) elif p == 2: return True _lowercase : Optional[Any] = 4 _lowercase : Tuple = (1 << p) - 1 for _ in range(p - 2 ): _lowercase : Union[str, Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
66
from __future__ import annotations UpperCamelCase = tuple[int, int, int] UpperCamelCase = tuple[str, str, str] # used alphabet -------------------------- # from string.ascii_uppercase UpperCamelCase = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" # -------------------------- default selection -------------------------- # rotors -------------------------- UpperCamelCase = "EGZWVONAHDCLFQMSIPJBYUKXTR" UpperCamelCase = "FOBHMDKEXQNRAULPGSJVTYICZW" UpperCamelCase = "ZJXESIUQLHAVRMDOYGTNFWPBKC" # reflector -------------------------- UpperCamelCase = { "A": "N", "N": "A", "B": "O", "O": "B", "C": "P", "P": "C", "D": "Q", "Q": "D", "E": "R", "R": "E", "F": "S", "S": "F", "G": "T", "T": "G", "H": "U", "U": "H", "I": "V", "V": "I", "J": "W", "W": "J", "K": "X", "X": "K", "L": "Y", "Y": "L", "M": "Z", "Z": "M", } # -------------------------- extra rotors -------------------------- UpperCamelCase = "RMDJXFUWGISLHVTCQNKYPBEZOA" UpperCamelCase = "SGLCPQWZHKXAREONTFBVIYJUDM" UpperCamelCase = "HVSICLTYKQUBXDWAJZOMFGPREN" UpperCamelCase = "RZWQHFMVDBKICJLNTUXAGYPSOE" UpperCamelCase = "LFKIJODBEGAMQPXVUHYSTCZRWN" UpperCamelCase = "KOAEGVDHXPQZMLFTYWJNBRCIUS" def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> tuple[RotorPositionT, RotorSelectionT, dict[str, str]]: # Checks if there are 3 unique rotors if (unique_rotsel := len(set(SCREAMING_SNAKE_CASE ) )) < 3: _lowercase : Optional[int] = F"""Please use 3 unique rotors (not {unique_rotsel})""" raise Exception(SCREAMING_SNAKE_CASE ) # Checks if rotor positions are valid _lowercase , _lowercase , _lowercase : int = rotpos if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : Dict = F"""First rotor position is not within range of 1..26 ({rotorposa}""" raise ValueError(SCREAMING_SNAKE_CASE ) if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : int = F"""Second rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(SCREAMING_SNAKE_CASE ) if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : str = F"""Third rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(SCREAMING_SNAKE_CASE ) # Validates string and returns dict _lowercase : Tuple = _plugboard(SCREAMING_SNAKE_CASE ) return rotpos, rotsel, pbdict def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> dict[str, str]: # tests the input string if it # a) is type string # b) has even length (so pairs can be made) if not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): _lowercase : Optional[int] = F"""Plugboard setting isn't type string ({type(SCREAMING_SNAKE_CASE )})""" raise TypeError(SCREAMING_SNAKE_CASE ) elif len(SCREAMING_SNAKE_CASE ) % 2 != 0: _lowercase : Optional[int] = F"""Odd number of symbols ({len(SCREAMING_SNAKE_CASE )})""" raise Exception(SCREAMING_SNAKE_CASE ) elif pbstring == "": return {} pbstring.replace(' ' , '' ) # Checks if all characters are unique _lowercase : Dict = set() for i in pbstring: if i not in abc: _lowercase : str = F"""'{i}' not in list of symbols""" raise Exception(SCREAMING_SNAKE_CASE ) elif i in tmppbl: _lowercase : int = F"""Duplicate symbol ({i})""" raise Exception(SCREAMING_SNAKE_CASE ) else: tmppbl.add(SCREAMING_SNAKE_CASE ) del tmppbl # Created the dictionary _lowercase : Optional[Any] = {} for j in range(0 , len(SCREAMING_SNAKE_CASE ) - 1 , 2 ): _lowercase : Dict = pbstring[j + 1] _lowercase : Union[str, Any] = pbstring[j] return pb def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = (rotora, rotora, rotora) , SCREAMING_SNAKE_CASE = "" , ) -> str: _lowercase : List[str] = text.upper() _lowercase , _lowercase , _lowercase : List[str] = _validator( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , plugb.upper() ) _lowercase , _lowercase , _lowercase : Optional[int] = rotor_position _lowercase , _lowercase , _lowercase : Union[str, Any] = rotor_selection rotorposa -= 1 rotorposa -= 1 rotorposa -= 1 _lowercase : Optional[int] = [] # encryption/decryption process -------------------------- for symbol in text: if symbol in abc: # 1st plugboard -------------------------- if symbol in plugboard: _lowercase : Dict = plugboard[symbol] # rotor ra -------------------------- _lowercase : Optional[Any] = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : Union[str, Any] = rotora[index % len(SCREAMING_SNAKE_CASE )] # rotor rb -------------------------- _lowercase : Tuple = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : str = rotora[index % len(SCREAMING_SNAKE_CASE )] # rotor rc -------------------------- _lowercase : List[Any] = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : List[str] = rotora[index % len(SCREAMING_SNAKE_CASE )] # reflector -------------------------- # this is the reason you don't need another machine to decipher _lowercase : List[str] = reflector[symbol] # 2nd rotors _lowercase : List[str] = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] _lowercase : Tuple = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] _lowercase : Dict = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] # 2nd plugboard if symbol in plugboard: _lowercase : int = plugboard[symbol] # moves/resets rotor positions rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : Any = 0 rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : int = 0 rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : Any = 0 # else: # pass # Error could be also raised # raise ValueError( # 'Invalid symbol('+repr(symbol)+')') result.append(SCREAMING_SNAKE_CASE ) return "".join(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": UpperCamelCase = "This is my Python script that emulates the Enigma machine from WWII." UpperCamelCase = (1, 1, 1) UpperCamelCase = "pictures" UpperCamelCase = (rotora, rotora, rotora) UpperCamelCase = enigma(message, rotor_pos, rotor_sel, pb) print("Encrypted message:", en) print("Decrypted message:", enigma(en, rotor_pos, rotor_sel, pb))
66
1
import os from huggingface_hub.constants import HUGGINGFACE_HUB_CACHE, hf_cache_home UpperCamelCase = HUGGINGFACE_HUB_CACHE UpperCamelCase = "config.json" UpperCamelCase = "diffusion_pytorch_model.bin" UpperCamelCase = "diffusion_flax_model.msgpack" UpperCamelCase = "model.onnx" UpperCamelCase = "diffusion_pytorch_model.safetensors" UpperCamelCase = "weights.pb" UpperCamelCase = "https://huggingface.co" UpperCamelCase = default_cache_path UpperCamelCase = "diffusers_modules" UpperCamelCase = os.getenv("HF_MODULES_CACHE", os.path.join(hf_cache_home, "modules")) UpperCamelCase = ["fp16", "non-ema"] UpperCamelCase = ".self_attn"
66
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = {"configuration_glpn": ["GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP", "GLPNConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["GLPNFeatureExtractor"] UpperCamelCase = ["GLPNImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "GLPN_PRETRAINED_MODEL_ARCHIVE_LIST", "GLPNForDepthEstimation", "GLPNLayer", "GLPNModel", "GLPNPreTrainedModel", ] if TYPE_CHECKING: from .configuration_glpn import GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP, GLPNConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_glpn import GLPNFeatureExtractor from .image_processing_glpn import GLPNImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_glpn import ( GLPN_PRETRAINED_MODEL_ARCHIVE_LIST, GLPNForDepthEstimation, GLPNLayer, GLPNModel, GLPNPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
1
from __future__ import annotations UpperCamelCase = list[list[int]] # assigning initial values to the grid UpperCamelCase = [ [3, 0, 6, 5, 0, 8, 4, 0, 0], [5, 2, 0, 0, 0, 0, 0, 0, 0], [0, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] # a grid with no solution UpperCamelCase = [ [5, 0, 6, 5, 0, 8, 4, 0, 3], [5, 2, 0, 0, 0, 0, 0, 0, 2], [1, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> bool: for i in range(9 ): if grid[row][i] == n or grid[i][column] == n: return False for i in range(3 ): for j in range(3 ): if grid[(row - row % 3) + i][(column - column % 3) + j] == n: return False return True def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> tuple[int, int] | None: for i in range(9 ): for j in range(9 ): if grid[i][j] == 0: return i, j return None def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> Matrix | None: if location := find_empty_location(SCREAMING_SNAKE_CASE ): _lowercase , _lowercase : Any = location else: # If the location is ``None``, then the grid is solved. return grid for digit in range(1 , 10 ): if is_safe(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): _lowercase : Optional[Any] = digit if sudoku(SCREAMING_SNAKE_CASE ) is not None: return grid _lowercase : List[Any] = 0 return None def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> None: for row in grid: for cell in row: print(SCREAMING_SNAKE_CASE , end=' ' ) print() if __name__ == "__main__": # make a copy of grid so that you can compare with the unmodified grid for example_grid in (initial_grid, no_solution): print("\nExample grid:\n" + "=" * 20) print_solution(example_grid) print("\nExample grid solution:") UpperCamelCase = sudoku(example_grid) if solution is not None: print_solution(solution) else: print("Cannot find a solution.")
66
import unittest from transformers import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device if is_torch_available(): import torch from transformers import AutoModelForImageClassification if is_vision_available(): from transformers import AutoImageProcessor @require_torch @require_vision class lowerCAmelCase_ ( unittest.TestCase ): @slow def __a ( self ): _lowercase : List[Any] = AutoImageProcessor.from_pretrained('microsoft/dit-base-finetuned-rvlcdip' ) _lowercase : List[Any] = AutoModelForImageClassification.from_pretrained('microsoft/dit-base-finetuned-rvlcdip' ) model.to(_lowerCAmelCase ) from datasets import load_dataset _lowercase : Union[str, Any] = load_dataset('nielsr/rvlcdip-demo' ) _lowercase : Any = dataset['train'][0]['image'].convert('RGB' ) _lowercase : List[str] = image_processor(_lowerCAmelCase , return_tensors='pt' ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): _lowercase : Dict = model(**_lowerCAmelCase ) _lowercase : Any = outputs.logits _lowercase : str = torch.Size((1, 1_6) ) self.assertEqual(logits.shape , _lowerCAmelCase ) _lowercase : Union[str, Any] = torch.tensor( [-0.41_58, -0.40_92, -0.43_47] , device=_lowerCAmelCase , dtype=torch.float , ) self.assertTrue(torch.allclose(logits[0, :3] , _lowerCAmelCase , atol=1E-4 ) )
66
1
def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> list[list]: _lowercase : Optional[int] = current_set.copy() for row_index, row in enumerate(SCREAMING_SNAKE_CASE ): _lowercase : str = row[0] for column_index, column in enumerate(SCREAMING_SNAKE_CASE ): if magnitude == 0: _lowercase : int = column continue _lowercase : Optional[Any] = column / magnitude # Subtract to cancel term _lowercase : List[Any] = current_set[0] _lowercase : Tuple = [first_row] _lowercase : str = current_set[1::] for row in current_set: _lowercase : Any = [] # If first term is 0, it is already in form we want, so we preserve it if row[0] == 0: final_set.append(SCREAMING_SNAKE_CASE ) continue for column_index in range(len(SCREAMING_SNAKE_CASE ) ): temp_row.append(first_row[column_index] - row[column_index] ) final_set.append(SCREAMING_SNAKE_CASE ) # Create next recursion iteration set if len(final_set[0] ) != 3: _lowercase : Optional[int] = final_set[0] _lowercase : Optional[int] = [] _lowercase : Any = [] for row in final_set[1::]: current_first_column.append(row[0] ) next_iteration.append(row[1::] ) _lowercase : int = simplify(SCREAMING_SNAKE_CASE ) for i in range(len(SCREAMING_SNAKE_CASE ) ): resultant[i].insert(0 , current_first_column[i] ) resultant.insert(0 , SCREAMING_SNAKE_CASE ) _lowercase : str = resultant return final_set def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> list: if len(SCREAMING_SNAKE_CASE ) == 0: raise IndexError('solve_simultaneous() requires n lists of length n+1' ) _lowercase : Dict = len(SCREAMING_SNAKE_CASE ) + 1 if any(len(SCREAMING_SNAKE_CASE ) != _length for item in equations ): raise IndexError('solve_simultaneous() requires n lists of length n+1' ) for row in equations: if any(not isinstance(SCREAMING_SNAKE_CASE , (int, float) ) for column in row ): raise ValueError('solve_simultaneous() requires lists of integers' ) if len(SCREAMING_SNAKE_CASE ) == 1: return [equations[0][-1] / equations[0][0]] _lowercase : Any = equations.copy() if any(0 in row for row in data_set ): _lowercase : int = data_set.copy() _lowercase : Optional[int] = [] for row_index, row in enumerate(SCREAMING_SNAKE_CASE ): if 0 not in row: _lowercase : Any = data_set.pop(SCREAMING_SNAKE_CASE ) break if not full_row: raise ValueError('solve_simultaneous() requires at least 1 full equation' ) data_set.insert(0 , SCREAMING_SNAKE_CASE ) _lowercase : Tuple = data_set.copy() _lowercase : List[str] = simplify(SCREAMING_SNAKE_CASE ) _lowercase : Any = simplified[::-1] _lowercase : list = [] for row in simplified: _lowercase : Union[str, Any] = row[-1] if not solutions: if row[-2] == 0: solutions.append(0 ) continue solutions.append(current_solution / row[-2] ) continue _lowercase : Optional[Any] = row.copy()[: len(SCREAMING_SNAKE_CASE ) - 1 :] while temp_row[0] == 0: temp_row.pop(0 ) if len(SCREAMING_SNAKE_CASE ) == 0: solutions.append(0 ) continue _lowercase : str = temp_row[1::] _lowercase : str = temp_row[::-1] for column_index, column in enumerate(SCREAMING_SNAKE_CASE ): current_solution -= column * solutions[column_index] solutions.append(SCREAMING_SNAKE_CASE ) _lowercase : int = [] for item in solutions: final.append(float(round(SCREAMING_SNAKE_CASE , 5 ) ) ) return final[::-1] if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase = [ [2, 1, 1, 1, 1, 4], [1, 2, 1, 1, 1, 5], [1, 1, 2, 1, 1, 6], [1, 1, 1, 2, 1, 7], [1, 1, 1, 1, 2, 8], ] print(solve_simultaneous(eq)) print(solve_simultaneous([[4, 2]]))
66
from PIL import Image def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Image: def brightness(SCREAMING_SNAKE_CASE ) -> float: return 128 + level + (c - 128) if not -255.0 <= level <= 255.0: raise ValueError('level must be between -255.0 (black) and 255.0 (white)' ) return img.point(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": # Load image with Image.open("image_data/lena.jpg") as img: # Change brightness to 100 UpperCamelCase = change_brightness(img, 100) brigt_img.save("image_data/lena_brightness.png", format="png")
66
1
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> float: if principal <= 0: raise Exception('Principal borrowed must be > 0' ) if rate_per_annum < 0: raise Exception('Rate of interest must be >= 0' ) if years_to_repay <= 0 or not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): raise Exception('Years to repay must be an integer > 0' ) # Yearly rate is divided by 12 to get monthly rate _lowercase : str = rate_per_annum / 12 # Years to repay is multiplied by 12 to get number of payments as payment is monthly _lowercase : List[Any] = years_to_repay * 12 return ( principal * rate_per_month * (1 + rate_per_month) ** number_of_payments / ((1 + rate_per_month) ** number_of_payments - 1) ) if __name__ == "__main__": import doctest doctest.testmod()
66
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): _lowercase : List[Any] = torch.nn.Linear(1_0 , 1_0 ) _lowercase : Any = torch.optim.SGD(model.parameters() , 0.1 ) _lowercase : str = Accelerator() _lowercase : Any = accelerator.prepare(_lowerCAmelCase ) try: pickle.loads(pickle.dumps(_lowerCAmelCase ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
66
1
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING UpperCamelCase = logging.get_logger(__name__) @add_end_docstrings(__snake_case ) class lowerCAmelCase_ ( __snake_case ): def __init__( self , **_lowerCAmelCase ): super().__init__(**_lowerCAmelCase ) if self.framework == "tf": raise ValueError(F"""The {self.__class__} is only available in PyTorch.""" ) requires_backends(self , 'vision' ) self.check_model_type(_lowerCAmelCase ) def __call__( self , _lowerCAmelCase , _lowerCAmelCase = None , **_lowerCAmelCase , ): if "text_queries" in kwargs: _lowercase : Union[str, Any] = kwargs.pop('text_queries' ) if isinstance(_lowerCAmelCase , (str, Image.Image) ): _lowercase : List[Any] = {'image': image, 'candidate_labels': candidate_labels} else: _lowercase : Union[str, Any] = image _lowercase : Dict = super().__call__(_lowerCAmelCase , **_lowerCAmelCase ) return results def __a ( self , **_lowerCAmelCase ): _lowercase : Optional[int] = {} if "threshold" in kwargs: _lowercase : Dict = kwargs['threshold'] if "top_k" in kwargs: _lowercase : int = kwargs['top_k'] return {}, {}, postprocess_params def __a ( self , _lowerCAmelCase ): _lowercase : Dict = load_image(inputs['image'] ) _lowercase : Any = inputs['candidate_labels'] if isinstance(_lowerCAmelCase , _lowerCAmelCase ): _lowercase : Tuple = candidate_labels.split(',' ) _lowercase : int = torch.tensor([[image.height, image.width]] , dtype=torch.intaa ) for i, candidate_label in enumerate(_lowerCAmelCase ): _lowercase : Union[str, Any] = self.tokenizer(_lowerCAmelCase , return_tensors=self.framework ) _lowercase : Union[str, Any] = self.image_processor(_lowerCAmelCase , return_tensors=self.framework ) yield { "is_last": i == len(_lowerCAmelCase ) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def __a ( self , _lowerCAmelCase ): _lowercase : Dict = model_inputs.pop('target_size' ) _lowercase : Optional[Any] = model_inputs.pop('candidate_label' ) _lowercase : int = model_inputs.pop('is_last' ) _lowercase : Dict = self.model(**_lowerCAmelCase ) _lowercase : Dict = {'target_size': target_size, 'candidate_label': candidate_label, 'is_last': is_last, **outputs} return model_outputs def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0.1 , _lowerCAmelCase=None ): _lowercase : List[Any] = [] for model_output in model_outputs: _lowercase : List[str] = model_output['candidate_label'] _lowercase : Optional[int] = BaseModelOutput(_lowerCAmelCase ) _lowercase : List[str] = self.image_processor.post_process_object_detection( outputs=_lowerCAmelCase , threshold=_lowerCAmelCase , target_sizes=model_output['target_size'] )[0] for index in outputs["scores"].nonzero(): _lowercase : Tuple = outputs['scores'][index].item() _lowercase : Any = self._get_bounding_box(outputs['boxes'][index][0] ) _lowercase : Dict = {'score': score, 'label': label, 'box': box} results.append(_lowerCAmelCase ) _lowercase : List[str] = sorted(_lowerCAmelCase , key=lambda _lowerCAmelCase : x["score"] , reverse=_lowerCAmelCase ) if top_k: _lowercase : List[Any] = results[:top_k] return results def __a ( self , _lowerCAmelCase ): if self.framework != "pt": raise ValueError('The ZeroShotObjectDetectionPipeline is only available in PyTorch.' ) _lowercase , _lowercase , _lowercase , _lowercase : Optional[Any] = box.int().tolist() _lowercase : Dict = { 'xmin': xmin, 'ymin': ymin, 'xmax': xmax, 'ymax': ymax, } return bbox
66
import requests from bsa import BeautifulSoup def __magic_name__ ( SCREAMING_SNAKE_CASE = "AAPL" ) -> str: _lowercase : str = F"""https://in.finance.yahoo.com/quote/{symbol}?s={symbol}""" _lowercase : int = BeautifulSoup(requests.get(SCREAMING_SNAKE_CASE ).text , 'html.parser' ) _lowercase : List[str] = 'My(6px) Pos(r) smartphone_Mt(6px)' return soup.find('div' , class_=class_ ).find('span' ).text if __name__ == "__main__": for symbol in "AAPL AMZN IBM GOOG MSFT ORCL".split(): print(f'''Current {symbol:<4} stock price is {stock_price(symbol):>8}''')
66
1
import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=1_024 , SCREAMING_SNAKE_CASE=1_024 , SCREAMING_SNAKE_CASE=False , **SCREAMING_SNAKE_CASE ) -> int: _lowercase : Optional[Any] = AutoTokenizer.from_pretrained(SCREAMING_SNAKE_CASE ) _lowercase : List[str] = SeqaSeqDataset(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , type_path='train' , **SCREAMING_SNAKE_CASE ) _lowercase : Dict = tok.pad_token_id def get_lens(SCREAMING_SNAKE_CASE ): _lowercase : Any = tqdm( DataLoader(SCREAMING_SNAKE_CASE , batch_size=512 , num_workers=8 , shuffle=SCREAMING_SNAKE_CASE , collate_fn=ds.collate_fn ) , desc=str(ds.len_file ) , ) _lowercase : Union[str, Any] = [] for batch in dl: _lowercase : Dict = batch['input_ids'].ne(SCREAMING_SNAKE_CASE ).sum(1 ).tolist() _lowercase : Union[str, Any] = batch['labels'].ne(SCREAMING_SNAKE_CASE ).sum(1 ).tolist() if consider_target: for src, tgt in zip(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): max_lens.append(max(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) else: max_lens.extend(SCREAMING_SNAKE_CASE ) return max_lens _lowercase : str = get_lens(SCREAMING_SNAKE_CASE ) _lowercase : Optional[int] = SeqaSeqDataset(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , type_path='val' , **SCREAMING_SNAKE_CASE ) _lowercase : List[Any] = get_lens(SCREAMING_SNAKE_CASE ) pickle_save(SCREAMING_SNAKE_CASE , train_ds.len_file ) pickle_save(SCREAMING_SNAKE_CASE , val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
66
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = { "configuration_poolformer": [ "POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "PoolFormerConfig", "PoolFormerOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["PoolFormerFeatureExtractor"] UpperCamelCase = ["PoolFormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "PoolFormerForImageClassification", "PoolFormerModel", "PoolFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure)
66
1
from PIL import Image def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Image: def brightness(SCREAMING_SNAKE_CASE ) -> float: return 128 + level + (c - 128) if not -255.0 <= level <= 255.0: raise ValueError('level must be between -255.0 (black) and 255.0 (white)' ) return img.point(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": # Load image with Image.open("image_data/lena.jpg") as img: # Change brightness to 100 UpperCamelCase = change_brightness(img, 100) brigt_img.save("image_data/lena_brightness.png", format="png")
66
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : Dict = LayoutLMTokenizer _UpperCamelCase : Union[str, Any] = LayoutLMTokenizerFast _UpperCamelCase : int = True _UpperCamelCase : Tuple = True def __a ( self ): super().setUp() _lowercase : Union[str, Any] = [ '[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] _lowercase : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def __a ( self , **_lowerCAmelCase ): return LayoutLMTokenizer.from_pretrained(self.tmpdirname , **_lowerCAmelCase ) def __a ( self , _lowerCAmelCase ): _lowercase : str = 'UNwant\u00E9d,running' _lowercase : List[Any] = 'unwanted, running' return input_text, output_text def __a ( self ): _lowercase : Dict = self.tokenizer_class(self.vocab_file ) _lowercase : Dict = tokenizer.tokenize('UNwant\u00E9d,running' ) self.assertListEqual(_lowerCAmelCase , ['un', '##want', '##ed', ',', 'runn', '##ing'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [7, 4, 5, 1_0, 8, 9] ) def __a ( self ): pass
66
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available UpperCamelCase = { "configuration_gpt_neo": ["GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTNeoConfig", "GPTNeoOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTNeoForCausalLM", "GPTNeoForQuestionAnswering", "GPTNeoForSequenceClassification", "GPTNeoForTokenClassification", "GPTNeoModel", "GPTNeoPreTrainedModel", "load_tf_weights_in_gpt_neo", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "FlaxGPTNeoForCausalLM", "FlaxGPTNeoModel", "FlaxGPTNeoPreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_neo import GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoConfig, GPTNeoOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neo import ( GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoForCausalLM, GPTNeoForQuestionAnswering, GPTNeoForSequenceClassification, GPTNeoForTokenClassification, GPTNeoModel, GPTNeoPreTrainedModel, load_tf_weights_in_gpt_neo, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_gpt_neo import FlaxGPTNeoForCausalLM, FlaxGPTNeoModel, FlaxGPTNeoPreTrainedModel else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : str = ShapEPipeline _UpperCamelCase : Any = ["prompt"] _UpperCamelCase : int = ["prompt"] _UpperCamelCase : Union[str, Any] = [ "num_images_per_prompt", "num_inference_steps", "generator", "latents", "guidance_scale", "frame_size", "output_type", "return_dict", ] _UpperCamelCase : Optional[Any] = False @property def __a ( self ): return 3_2 @property def __a ( self ): return 3_2 @property def __a ( self ): return self.time_input_dim * 4 @property def __a ( self ): return 8 @property def __a ( self ): _lowercase : Any = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) return tokenizer @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) return CLIPTextModelWithProjection(_lowerCAmelCase ) @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = { 'num_attention_heads': 2, 'attention_head_dim': 1_6, 'embedding_dim': self.time_input_dim, 'num_embeddings': 3_2, 'embedding_proj_dim': self.text_embedder_hidden_size, 'time_embed_dim': self.time_embed_dim, 'num_layers': 1, 'clip_embed_dim': self.time_input_dim * 2, 'additional_embeddings': 0, 'time_embed_act_fn': 'gelu', 'norm_in_type': 'layer', 'encoder_hid_proj_type': None, 'added_emb_type': None, } _lowercase : Optional[Any] = PriorTransformer(**_lowerCAmelCase ) return model @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = { 'param_shapes': ( (self.renderer_dim, 9_3), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), 'd_latent': self.time_input_dim, 'd_hidden': self.renderer_dim, 'n_output': 1_2, 'background': ( 0.1, 0.1, 0.1, ), } _lowercase : List[Any] = ShapERenderer(**_lowerCAmelCase ) return model def __a ( self ): _lowercase : Optional[Any] = self.dummy_prior _lowercase : Dict = self.dummy_text_encoder _lowercase : List[str] = self.dummy_tokenizer _lowercase : Union[str, Any] = self.dummy_renderer _lowercase : List[str] = HeunDiscreteScheduler( beta_schedule='exp' , num_train_timesteps=1_0_2_4 , prediction_type='sample' , use_karras_sigmas=_lowerCAmelCase , clip_sample=_lowerCAmelCase , clip_sample_range=1.0 , ) _lowercase : List[str] = { 'prior': prior, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'renderer': renderer, 'scheduler': scheduler, } return components def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : Optional[Any] = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Optional[int] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : List[Any] = { 'prompt': 'horse', 'generator': generator, 'num_inference_steps': 1, 'frame_size': 3_2, 'output_type': 'np', } return inputs def __a ( self ): _lowercase : Optional[int] = 'cpu' _lowercase : List[Any] = self.get_dummy_components() _lowercase : Tuple = self.pipeline_class(**_lowerCAmelCase ) _lowercase : Union[str, Any] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Union[str, Any] = pipe(**self.get_dummy_inputs(_lowerCAmelCase ) ) _lowercase : str = output.images[0] _lowercase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (2_0, 3_2, 3_2, 3) _lowercase : str = np.array( [ 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __a ( self ): # NOTE: Larger batch sizes cause this test to timeout, only test on smaller batches self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def __a ( self ): _lowercase : List[Any] = torch_device == 'cpu' _lowercase : Any = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=_lowerCAmelCase , relax_max_difference=_lowerCAmelCase , ) def __a ( self ): _lowercase : Union[str, Any] = self.get_dummy_components() _lowercase : Optional[int] = self.pipeline_class(**_lowerCAmelCase ) _lowercase : Any = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : str = 1 _lowercase : Optional[int] = 2 _lowercase : List[str] = self.get_dummy_inputs(_lowerCAmelCase ) for key in inputs.keys(): if key in self.batch_params: _lowercase : int = batch_size * [inputs[key]] _lowercase : Optional[int] = pipe(**_lowerCAmelCase , num_images_per_prompt=_lowerCAmelCase )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ): _lowercase : List[str] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/shap_e/test_shap_e_np_out.npy' ) _lowercase : Any = ShapEPipeline.from_pretrained('openai/shap-e' ) _lowercase : List[str] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Tuple = torch.Generator(device=_lowerCAmelCase ).manual_seed(0 ) _lowercase : int = pipe( 'a shark' , generator=_lowerCAmelCase , guidance_scale=15.0 , num_inference_steps=6_4 , frame_size=6_4 , output_type='np' , ).images[0] assert images.shape == (2_0, 6_4, 6_4, 3) assert_mean_pixel_difference(_lowerCAmelCase , _lowerCAmelCase )
66
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCamelCase = { "configuration_groupvit": [ "GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "GroupViTConfig", "GroupViTOnnxConfig", "GroupViTTextConfig", "GroupViTVisionConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "GroupViTModel", "GroupViTPreTrainedModel", "GroupViTTextModel", "GroupViTVisionModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFGroupViTModel", "TFGroupViTPreTrainedModel", "TFGroupViTTextModel", "TFGroupViTVisionModel", ] if TYPE_CHECKING: from .configuration_groupvit import ( GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GroupViTConfig, GroupViTOnnxConfig, GroupViTTextConfig, GroupViTVisionConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_groupvit import ( GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, GroupViTModel, GroupViTPreTrainedModel, GroupViTTextModel, GroupViTVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_groupvit import ( TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFGroupViTModel, TFGroupViTPreTrainedModel, TFGroupViTTextModel, TFGroupViTVisionModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
import sys UpperCamelCase = ( "73167176531330624919225119674426574742355349194934" "96983520312774506326239578318016984801869478851843" "85861560789112949495459501737958331952853208805511" "12540698747158523863050715693290963295227443043557" "66896648950445244523161731856403098711121722383113" "62229893423380308135336276614282806444486645238749" "30358907296290491560440772390713810515859307960866" "70172427121883998797908792274921901699720888093776" "65727333001053367881220235421809751254540594752243" "52584907711670556013604839586446706324415722155397" "53697817977846174064955149290862569321978468622482" "83972241375657056057490261407972968652414535100474" "82166370484403199890008895243450658541227588666881" "16427171479924442928230863465674813919123162824586" "17866458359124566529476545682848912883142607690042" "24219022671055626321111109370544217506941658960408" "07198403850962455444362981230987879927244284909188" "84580156166097919133875499200524063689912560717606" "05886116467109405077541002256983155200055935729725" "71636269561882670428252483600823257530420752963450" ) def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : List[Any] = 1 for digit in s: product *= int(SCREAMING_SNAKE_CASE ) return product def __magic_name__ ( SCREAMING_SNAKE_CASE = N ) -> int: _lowercase : Dict = -sys.maxsize - 1 _lowercase : Tuple = n[:13] _lowercase : List[Any] = 13 while cur_index < len(SCREAMING_SNAKE_CASE ) - 13: if int(n[cur_index] ) >= int(substr[0] ): _lowercase : List[str] = substr[1:] + n[cur_index] cur_index += 1 else: _lowercase : str = max(SCREAMING_SNAKE_CASE , str_eval(SCREAMING_SNAKE_CASE ) ) _lowercase : Dict = n[cur_index : cur_index + 13] cur_index += 13 return largest_product if __name__ == "__main__": print(f'''{solution() = }''')
66
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = { "configuration_efficientnet": [ "EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP", "EfficientNetConfig", "EfficientNetOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["EfficientNetImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST", "EfficientNetForImageClassification", "EfficientNetModel", "EfficientNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_efficientnet import ( EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientNetConfig, EfficientNetOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientnet import EfficientNetImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientnet import ( EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientNetForImageClassification, EfficientNetModel, EfficientNetPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure)
66
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available UpperCamelCase = { "configuration_gpt_neo": ["GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTNeoConfig", "GPTNeoOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTNeoForCausalLM", "GPTNeoForQuestionAnswering", "GPTNeoForSequenceClassification", "GPTNeoForTokenClassification", "GPTNeoModel", "GPTNeoPreTrainedModel", "load_tf_weights_in_gpt_neo", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "FlaxGPTNeoForCausalLM", "FlaxGPTNeoModel", "FlaxGPTNeoPreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_neo import GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoConfig, GPTNeoOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neo import ( GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoForCausalLM, GPTNeoForQuestionAnswering, GPTNeoForSequenceClassification, GPTNeoForTokenClassification, GPTNeoModel, GPTNeoPreTrainedModel, load_tf_weights_in_gpt_neo, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_gpt_neo import FlaxGPTNeoForCausalLM, FlaxGPTNeoModel, FlaxGPTNeoPreTrainedModel else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
1
def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> str: return "".join([hex(SCREAMING_SNAKE_CASE )[2:].zfill(2 ).upper() for byte in list(SCREAMING_SNAKE_CASE )] ) def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> bytes: # Check data validity, following RFC3548 # https://www.ietf.org/rfc/rfc3548.txt if (len(SCREAMING_SNAKE_CASE ) % 2) != 0: raise ValueError( 'Base16 encoded data is invalid:\nData does not have an even number of hex digits.' ) # Check the character set - the standard base16 alphabet # is uppercase according to RFC3548 section 6 if not set(SCREAMING_SNAKE_CASE ) <= set('0123456789ABCDEF' ): raise ValueError( 'Base16 encoded data is invalid:\nData is not uppercase hex or it contains invalid characters.' ) # For every two hexadecimal digits (= a byte), turn it into an integer. # Then, string the result together into bytes, and return it. return bytes(int(data[i] + data[i + 1] , 16 ) for i in range(0 , len(SCREAMING_SNAKE_CASE ) , 2 ) ) if __name__ == "__main__": import doctest doctest.testmod()
66
from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : str = ["image_processor", "tokenizer"] _UpperCamelCase : Union[str, Any] = "AutoImageProcessor" _UpperCamelCase : Union[str, Any] = "AutoTokenizer" def __init__( self , _lowerCAmelCase , _lowerCAmelCase ): super().__init__(_lowerCAmelCase , _lowerCAmelCase ) _lowercase : Union[str, Any] = self.image_processor def __call__( self , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=None , **_lowerCAmelCase ): if text is None and images is None: raise ValueError('You have to specify either text or images. Both cannot be none.' ) if text is not None: _lowercase : Dict = self.tokenizer(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if images is not None: _lowercase : Union[str, Any] = self.image_processor(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if text is not None and images is not None: _lowercase : Optional[Any] = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_lowerCAmelCase ) , tensor_type=_lowerCAmelCase ) def __a ( self , *_lowerCAmelCase , **_lowerCAmelCase ): return self.tokenizer.batch_decode(*_lowerCAmelCase , **_lowerCAmelCase ) def __a ( self , *_lowerCAmelCase , **_lowerCAmelCase ): return self.tokenizer.decode(*_lowerCAmelCase , **_lowerCAmelCase ) @property def __a ( self ): return ["input_ids", "attention_mask", "pixel_values"]
66
1
from __future__ import annotations from decimal import Decimal from numpy import array def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> list[list[float]]: _lowercase : List[str] = Decimal # Check if the provided matrix has 2 rows and 2 columns # since this implementation only works for 2x2 matrices if len(SCREAMING_SNAKE_CASE ) == 2 and len(matrix[0] ) == 2 and len(matrix[1] ) == 2: # Calculate the determinant of the matrix _lowercase : List[Any] = float( d(matrix[0][0] ) * d(matrix[1][1] ) - d(matrix[1][0] ) * d(matrix[0][1] ) ) if determinant == 0: raise ValueError('This matrix has no inverse.' ) # Creates a copy of the matrix with swapped positions of the elements _lowercase : Optional[Any] = [[0.0, 0.0], [0.0, 0.0]] _lowercase , _lowercase : str = matrix[1][1], matrix[0][0] _lowercase , _lowercase : Optional[int] = -matrix[1][0], -matrix[0][1] # Calculate the inverse of the matrix return [ [(float(d(SCREAMING_SNAKE_CASE ) ) / determinant) or 0.0 for n in row] for row in swapped_matrix ] elif ( len(SCREAMING_SNAKE_CASE ) == 3 and len(matrix[0] ) == 3 and len(matrix[1] ) == 3 and len(matrix[2] ) == 3 ): # Calculate the determinant of the matrix using Sarrus rule _lowercase : str = float( ( (d(matrix[0][0] ) * d(matrix[1][1] ) * d(matrix[2][2] )) + (d(matrix[0][1] ) * d(matrix[1][2] ) * d(matrix[2][0] )) + (d(matrix[0][2] ) * d(matrix[1][0] ) * d(matrix[2][1] )) ) - ( (d(matrix[0][2] ) * d(matrix[1][1] ) * d(matrix[2][0] )) + (d(matrix[0][1] ) * d(matrix[1][0] ) * d(matrix[2][2] )) + (d(matrix[0][0] ) * d(matrix[1][2] ) * d(matrix[2][1] )) ) ) if determinant == 0: raise ValueError('This matrix has no inverse.' ) # Creating cofactor matrix _lowercase : str = [ [d(0.0 ), d(0.0 ), d(0.0 )], [d(0.0 ), d(0.0 ), d(0.0 )], [d(0.0 ), d(0.0 ), d(0.0 )], ] _lowercase : Dict = (d(matrix[1][1] ) * d(matrix[2][2] )) - ( d(matrix[1][2] ) * d(matrix[2][1] ) ) _lowercase : List[str] = -( (d(matrix[1][0] ) * d(matrix[2][2] )) - (d(matrix[1][2] ) * d(matrix[2][0] )) ) _lowercase : Dict = (d(matrix[1][0] ) * d(matrix[2][1] )) - ( d(matrix[1][1] ) * d(matrix[2][0] ) ) _lowercase : Dict = -( (d(matrix[0][1] ) * d(matrix[2][2] )) - (d(matrix[0][2] ) * d(matrix[2][1] )) ) _lowercase : Tuple = (d(matrix[0][0] ) * d(matrix[2][2] )) - ( d(matrix[0][2] ) * d(matrix[2][0] ) ) _lowercase : Optional[Any] = -( (d(matrix[0][0] ) * d(matrix[2][1] )) - (d(matrix[0][1] ) * d(matrix[2][0] )) ) _lowercase : Any = (d(matrix[0][1] ) * d(matrix[1][2] )) - ( d(matrix[0][2] ) * d(matrix[1][1] ) ) _lowercase : Dict = -( (d(matrix[0][0] ) * d(matrix[1][2] )) - (d(matrix[0][2] ) * d(matrix[1][0] )) ) _lowercase : List[str] = (d(matrix[0][0] ) * d(matrix[1][1] )) - ( d(matrix[0][1] ) * d(matrix[1][0] ) ) # Transpose the cofactor matrix (Adjoint matrix) _lowercase : Tuple = array(SCREAMING_SNAKE_CASE ) for i in range(3 ): for j in range(3 ): _lowercase : str = cofactor_matrix[j][i] # Inverse of the matrix using the formula (1/determinant) * adjoint matrix _lowercase : List[Any] = array(SCREAMING_SNAKE_CASE ) for i in range(3 ): for j in range(3 ): inverse_matrix[i][j] /= d(SCREAMING_SNAKE_CASE ) # Calculate the inverse of the matrix return [[float(d(SCREAMING_SNAKE_CASE ) ) or 0.0 for n in row] for row in inverse_matrix] raise ValueError('Please provide a matrix of size 2x2 or 3x3.' )
66
from __future__ import annotations import math def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> list[int]: if num <= 0: _lowercase : List[str] = F"""{num}: Invalid input, please enter a positive integer.""" raise ValueError(SCREAMING_SNAKE_CASE ) _lowercase : Union[str, Any] = [True] * (num + 1) _lowercase : Union[str, Any] = [] _lowercase : Dict = 2 _lowercase : Union[str, Any] = int(math.sqrt(SCREAMING_SNAKE_CASE ) ) while start <= end: # If start is a prime if sieve[start] is True: prime.append(SCREAMING_SNAKE_CASE ) # Set multiples of start be False for i in range(start * start , num + 1 , SCREAMING_SNAKE_CASE ): if sieve[i] is True: _lowercase : str = False start += 1 for j in range(end + 1 , num + 1 ): if sieve[j] is True: prime.append(SCREAMING_SNAKE_CASE ) return prime if __name__ == "__main__": print(prime_sieve(int(input("Enter a positive integer: ").strip())))
66
1
import re import jax.numpy as jnp from flax.traverse_util import flatten_dict, unflatten_dict from jax.random import PRNGKey from ..utils import logging UpperCamelCase = logging.get_logger(__name__) def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> Optional[Any]: _lowercase : Optional[Any] = R'\w+[.]\d+' _lowercase : str = re.findall(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for pat in pats: _lowercase : int = key.replace(SCREAMING_SNAKE_CASE , '_'.join(pat.split('.' ) ) ) return key def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Optional[Any]: _lowercase : Dict = pt_tuple_key[:-1] + ('scale',) if ( any('norm' in str_ for str_ in pt_tuple_key ) and (pt_tuple_key[-1] == "bias") and (pt_tuple_key[:-1] + ("bias",) not in random_flax_state_dict) and (pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict) ): _lowercase : str = pt_tuple_key[:-1] + ('scale',) return renamed_pt_tuple_key, pt_tensor elif pt_tuple_key[-1] in ["weight", "gamma"] and pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict: _lowercase : Optional[Any] = pt_tuple_key[:-1] + ('scale',) return renamed_pt_tuple_key, pt_tensor # embedding if pt_tuple_key[-1] == "weight" and pt_tuple_key[:-1] + ("embedding",) in random_flax_state_dict: _lowercase : str = pt_tuple_key[:-1] + ('embedding',) return renamed_pt_tuple_key, pt_tensor # conv layer _lowercase : List[str] = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4: _lowercase : Any = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer _lowercase : List[str] = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight": _lowercase : Any = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight _lowercase : Any = pt_tuple_key[:-1] + ('weight',) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias _lowercase : List[Any] = pt_tuple_key[:-1] + ('bias',) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=42 ) -> Optional[int]: # Step 1: Convert pytorch tensor to numpy _lowercase : Tuple = {k: v.numpy() for k, v in pt_state_dict.items()} # Step 2: Since the model is stateless, get random Flax params _lowercase : int = flax_model.init_weights(PRNGKey(SCREAMING_SNAKE_CASE ) ) _lowercase : str = flatten_dict(SCREAMING_SNAKE_CASE ) _lowercase : Dict = {} # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): _lowercase : Optional[Any] = rename_key(SCREAMING_SNAKE_CASE ) _lowercase : Dict = tuple(renamed_pt_key.split('.' ) ) # Correctly rename weight parameters _lowercase , _lowercase : Optional[Any] = rename_key_and_reshape_tensor(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( F"""PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape """ F"""{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) # also add unexpected weight so that warning is thrown _lowercase : str = jnp.asarray(SCREAMING_SNAKE_CASE ) return unflatten_dict(SCREAMING_SNAKE_CASE )
66
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ConvNextConfig, SegformerImageProcessor, UperNetConfig, UperNetForSemanticSegmentation def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> List[Any]: _lowercase : int = 384 if "tiny" in model_name: _lowercase : Tuple = [3, 3, 9, 3] _lowercase : List[str] = [96, 192, 384, 768] if "small" in model_name: _lowercase : List[str] = [3, 3, 27, 3] _lowercase : Union[str, Any] = [96, 192, 384, 768] if "base" in model_name: _lowercase : List[Any] = [3, 3, 27, 3] _lowercase : Dict = [128, 256, 512, 1_024] _lowercase : Optional[int] = 512 if "large" in model_name: _lowercase : List[str] = [3, 3, 27, 3] _lowercase : List[Any] = [192, 384, 768, 1_536] _lowercase : Tuple = 768 if "xlarge" in model_name: _lowercase : str = [3, 3, 27, 3] _lowercase : List[str] = [256, 512, 1_024, 2_048] _lowercase : Tuple = 1_024 # set label information _lowercase : Dict = 150 _lowercase : Union[str, Any] = 'huggingface/label-files' _lowercase : str = 'ade20k-id2label.json' _lowercase : List[Any] = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , repo_type='dataset' ) , 'r' ) ) _lowercase : Dict = {int(SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} _lowercase : Tuple = {v: k for k, v in idalabel.items()} _lowercase : List[str] = ConvNextConfig( depths=SCREAMING_SNAKE_CASE , hidden_sizes=SCREAMING_SNAKE_CASE , out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) _lowercase : Union[str, Any] = UperNetConfig( backbone_config=SCREAMING_SNAKE_CASE , auxiliary_in_channels=SCREAMING_SNAKE_CASE , num_labels=SCREAMING_SNAKE_CASE , idalabel=SCREAMING_SNAKE_CASE , labelaid=SCREAMING_SNAKE_CASE , ) return config def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : Any = [] # fmt: off # stem rename_keys.append(('backbone.downsample_layers.0.0.weight', 'backbone.embeddings.patch_embeddings.weight') ) rename_keys.append(('backbone.downsample_layers.0.0.bias', 'backbone.embeddings.patch_embeddings.bias') ) rename_keys.append(('backbone.downsample_layers.0.1.weight', 'backbone.embeddings.layernorm.weight') ) rename_keys.append(('backbone.downsample_layers.0.1.bias', 'backbone.embeddings.layernorm.bias') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F"""backbone.stages.{i}.{j}.gamma""", F"""backbone.encoder.stages.{i}.layers.{j}.layer_scale_parameter""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.depthwise_conv.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.dwconv.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.depthwise_conv.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.dwconv.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.norm.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.layernorm.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.norm.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.layernorm.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv1.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv1.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv1.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv1.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv2.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv2.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv2.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv2.bias""") ) if i > 0: rename_keys.append((F"""backbone.downsample_layers.{i}.0.weight""", F"""backbone.encoder.stages.{i}.downsampling_layer.0.weight""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.0.bias""", F"""backbone.encoder.stages.{i}.downsampling_layer.0.bias""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.1.weight""", F"""backbone.encoder.stages.{i}.downsampling_layer.1.weight""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.1.bias""", F"""backbone.encoder.stages.{i}.downsampling_layer.1.bias""") ) rename_keys.append((F"""backbone.norm{i}.weight""", F"""backbone.hidden_states_norms.stage{i+1}.weight""") ) rename_keys.append((F"""backbone.norm{i}.bias""", F"""backbone.hidden_states_norms.stage{i+1}.bias""") ) # decode head rename_keys.extend( [ ('decode_head.conv_seg.weight', 'decode_head.classifier.weight'), ('decode_head.conv_seg.bias', 'decode_head.classifier.bias'), ('auxiliary_head.conv_seg.weight', 'auxiliary_head.classifier.weight'), ('auxiliary_head.conv_seg.bias', 'auxiliary_head.classifier.bias'), ] ) # fmt: on return rename_keys def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Optional[int]: _lowercase : Any = dct.pop(SCREAMING_SNAKE_CASE ) _lowercase : Any = val def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Any: _lowercase : List[Any] = { 'upernet-convnext-tiny': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_tiny_fp16_512x512_160k_ade20k/upernet_convnext_tiny_fp16_512x512_160k_ade20k_20220227_124553-cad485de.pth', 'upernet-convnext-small': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_small_fp16_512x512_160k_ade20k/upernet_convnext_small_fp16_512x512_160k_ade20k_20220227_131208-1b1e394f.pth', 'upernet-convnext-base': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_base_fp16_512x512_160k_ade20k/upernet_convnext_base_fp16_512x512_160k_ade20k_20220227_181227-02a24fc6.pth', 'upernet-convnext-large': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_large_fp16_640x640_160k_ade20k/upernet_convnext_large_fp16_640x640_160k_ade20k_20220226_040532-e57aa54d.pth', 'upernet-convnext-xlarge': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_xlarge_fp16_640x640_160k_ade20k/upernet_convnext_xlarge_fp16_640x640_160k_ade20k_20220226_080344-95fc38c2.pth', } _lowercase : Optional[int] = model_name_to_url[model_name] _lowercase : str = torch.hub.load_state_dict_from_url(SCREAMING_SNAKE_CASE , map_location='cpu' )['state_dict'] _lowercase : Optional[int] = get_upernet_config(SCREAMING_SNAKE_CASE ) _lowercase : Tuple = UperNetForSemanticSegmentation(SCREAMING_SNAKE_CASE ) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): _lowercase : List[Any] = state_dict.pop(SCREAMING_SNAKE_CASE ) if "bn" in key: _lowercase : Any = key.replace('bn' , 'batch_norm' ) _lowercase : Any = val # rename keys _lowercase : int = create_rename_keys(SCREAMING_SNAKE_CASE ) for src, dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) model.load_state_dict(SCREAMING_SNAKE_CASE ) # verify on image _lowercase : Union[str, Any] = 'https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg' _lowercase : Any = Image.open(requests.get(SCREAMING_SNAKE_CASE , stream=SCREAMING_SNAKE_CASE ).raw ).convert('RGB' ) _lowercase : Tuple = SegformerImageProcessor() _lowercase : Tuple = processor(SCREAMING_SNAKE_CASE , return_tensors='pt' ).pixel_values with torch.no_grad(): _lowercase : Dict = model(SCREAMING_SNAKE_CASE ) if model_name == "upernet-convnext-tiny": _lowercase : Dict = torch.tensor( [[-8.8110, -8.8110, -8.6521], [-8.8110, -8.8110, -8.6521], [-8.7746, -8.7746, -8.6130]] ) elif model_name == "upernet-convnext-small": _lowercase : Union[str, Any] = torch.tensor( [[-8.8236, -8.8236, -8.6771], [-8.8236, -8.8236, -8.6771], [-8.7638, -8.7638, -8.6240]] ) elif model_name == "upernet-convnext-base": _lowercase : Dict = torch.tensor( [[-8.8558, -8.8558, -8.6905], [-8.8558, -8.8558, -8.6905], [-8.7669, -8.7669, -8.6021]] ) elif model_name == "upernet-convnext-large": _lowercase : Optional[int] = torch.tensor( [[-8.6660, -8.6660, -8.6210], [-8.6660, -8.6660, -8.6210], [-8.6310, -8.6310, -8.5964]] ) elif model_name == "upernet-convnext-xlarge": _lowercase : str = torch.tensor( [[-8.4980, -8.4980, -8.3977], [-8.4980, -8.4980, -8.3977], [-8.4379, -8.4379, -8.3412]] ) print('Logits:' , outputs.logits[0, 0, :3, :3] ) assert torch.allclose(outputs.logits[0, 0, :3, :3] , SCREAMING_SNAKE_CASE , atol=1E-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(SCREAMING_SNAKE_CASE ) print(F"""Saving processor to {pytorch_dump_folder_path}""" ) processor.save_pretrained(SCREAMING_SNAKE_CASE ) if push_to_hub: print(F"""Pushing model and processor for {model_name} to hub""" ) model.push_to_hub(F"""openmmlab/{model_name}""" ) processor.push_to_hub(F"""openmmlab/{model_name}""" ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="upernet-convnext-tiny", type=str, choices=[f'''upernet-convnext-{size}''' for size in ["tiny", "small", "base", "large", "xlarge"]], help="Name of the ConvNext UperNet model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) UpperCamelCase = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
66
1
import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : int = args.pruning_method _lowercase : Optional[int] = args.threshold _lowercase : str = args.model_name_or_path.rstrip('/' ) _lowercase : Optional[Any] = args.target_model_path print(F"""Load fine-pruned model from {model_name_or_path}""" ) _lowercase : int = torch.load(os.path.join(SCREAMING_SNAKE_CASE , 'pytorch_model.bin' ) ) _lowercase : List[Any] = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: _lowercase : Optional[Any] = tensor print(F"""Copied layer {name}""" ) elif "classifier" in name or "qa_output" in name: _lowercase : Tuple = tensor print(F"""Copied layer {name}""" ) elif "bias" in name: _lowercase : List[Any] = tensor print(F"""Copied layer {name}""" ) else: if pruning_method == "magnitude": _lowercase : Optional[int] = MagnitudeBinarizer.apply(inputs=SCREAMING_SNAKE_CASE , threshold=SCREAMING_SNAKE_CASE ) _lowercase : Dict = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "topK": if "mask_scores" in name: continue _lowercase : List[Any] = name[:-6] _lowercase : Dict = model[F"""{prefix_}mask_scores"""] _lowercase : Any = TopKBinarizer.apply(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) _lowercase : List[Any] = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue _lowercase : Dict = name[:-6] _lowercase : Union[str, Any] = model[F"""{prefix_}mask_scores"""] _lowercase : Any = ThresholdBinarizer.apply(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) _lowercase : str = tensor * mask print(F"""Pruned layer {name}""" ) elif pruning_method == "l0": if "mask_scores" in name: continue _lowercase : Optional[Any] = name[:-6] _lowercase : Optional[int] = model[F"""{prefix_}mask_scores"""] _lowercase , _lowercase : List[str] = -0.1, 1.1 _lowercase : int = torch.sigmoid(SCREAMING_SNAKE_CASE ) _lowercase : Tuple = s * (r - l) + l _lowercase : Union[str, Any] = s_bar.clamp(min=0.0 , max=1.0 ) _lowercase : Optional[int] = tensor * mask print(F"""Pruned layer {name}""" ) else: raise ValueError('Unknown pruning method' ) if target_model_path is None: _lowercase : Tuple = os.path.join( os.path.dirname(SCREAMING_SNAKE_CASE ) , F"""bertarized_{os.path.basename(SCREAMING_SNAKE_CASE )}""" ) if not os.path.isdir(SCREAMING_SNAKE_CASE ): shutil.copytree(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) print(F"""\nCreated folder {target_model_path}""" ) torch.save(SCREAMING_SNAKE_CASE , os.path.join(SCREAMING_SNAKE_CASE , 'pytorch_model.bin' ) ) print('\nPruned model saved! See you later!' ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() parser.add_argument( "--pruning_method", choices=["l0", "magnitude", "topK", "sigmoied_threshold"], type=str, required=True, help=( "Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning," " sigmoied_threshold = Soft movement pruning)" ), ) parser.add_argument( "--threshold", type=float, required=False, help=( "For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model." "For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared." "Not needed for `l0`" ), ) parser.add_argument( "--model_name_or_path", type=str, required=True, help="Folder containing the model that was previously fine-pruned", ) parser.add_argument( "--target_model_path", default=None, type=str, required=False, help="Folder containing the model that was previously fine-pruned", ) UpperCamelCase = parser.parse_args() main(args)
66
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto.configuration_auto import CONFIG_MAPPING UpperCamelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : Any = "upernet" def __init__( self , _lowerCAmelCase=None , _lowerCAmelCase=5_1_2 , _lowerCAmelCase=0.02 , _lowerCAmelCase=[1, 2, 3, 6] , _lowerCAmelCase=True , _lowerCAmelCase=0.4 , _lowerCAmelCase=3_8_4 , _lowerCAmelCase=2_5_6 , _lowerCAmelCase=1 , _lowerCAmelCase=False , _lowerCAmelCase=2_5_5 , **_lowerCAmelCase , ): super().__init__(**_lowerCAmelCase ) if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.' ) _lowercase : Optional[Any] = CONFIG_MAPPING['resnet'](out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): _lowercase : List[Any] = backbone_config.get('model_type' ) _lowercase : str = CONFIG_MAPPING[backbone_model_type] _lowercase : Tuple = config_class.from_dict(_lowerCAmelCase ) _lowercase : Optional[Any] = backbone_config _lowercase : Any = hidden_size _lowercase : Any = initializer_range _lowercase : Tuple = pool_scales _lowercase : List[Any] = use_auxiliary_head _lowercase : Optional[Any] = auxiliary_loss_weight _lowercase : Any = auxiliary_in_channels _lowercase : Any = auxiliary_channels _lowercase : List[str] = auxiliary_num_convs _lowercase : List[str] = auxiliary_concat_input _lowercase : Tuple = loss_ignore_index def __a ( self ): _lowercase : str = copy.deepcopy(self.__dict__ ) _lowercase : Tuple = self.backbone_config.to_dict() _lowercase : int = self.__class__.model_type return output
66
1
from copy import deepcopy import torch import torch.nn.functional as F from torch.optim import AdamW from torch.optim.lr_scheduler import LambdaLR from torch.utils.data import DataLoader from accelerate.accelerator import Accelerator from accelerate.state import GradientState from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import DistributedType, is_torch_version, set_seed def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Union[str, Any]: for param, grad_param in zip(model_a.parameters() , model_b.parameters() ): if not param.requires_grad: continue if not did_step: # Grads should not be in sync assert ( torch.allclose(param.grad , grad_param.grad ) is False ), F"""Gradients in sync when they should not be at iteration {iteration}:\nmodel_a grad ({param.grad}) == model_b grad ({grad_param.grad})""" else: # Grads should be in sync assert ( torch.allclose(param.grad , grad_param.grad ) is True ), F"""Gradients not in sync when they should be at iteration {iteration}:\nmodel_a grad ({param.grad}) != model_b grad ({grad_param.grad})""" def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=True ) -> Tuple: model.train() _lowercase : List[str] = model(SCREAMING_SNAKE_CASE ) _lowercase : Union[str, Any] = F.mse_loss(SCREAMING_SNAKE_CASE , target.to(output.device ) ) if not do_backward: loss /= accelerator.gradient_accumulation_steps loss.backward() else: accelerator.backward(SCREAMING_SNAKE_CASE ) def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=False ) -> List[Any]: set_seed(42 ) _lowercase : List[str] = RegressionModel() _lowercase : Any = deepcopy(SCREAMING_SNAKE_CASE ) _lowercase : List[Any] = RegressionDataset(length=80 ) _lowercase : Any = DataLoader(SCREAMING_SNAKE_CASE , batch_size=16 ) model.to(accelerator.device ) if sched: _lowercase : int = AdamW(params=model.parameters() , lr=1E-3 ) _lowercase : List[str] = AdamW(params=ddp_model.parameters() , lr=1E-3 ) _lowercase : List[Any] = LambdaLR(SCREAMING_SNAKE_CASE , lr_lambda=lambda SCREAMING_SNAKE_CASE : epoch**0.65 ) _lowercase : Any = LambdaLR(SCREAMING_SNAKE_CASE , lr_lambda=lambda SCREAMING_SNAKE_CASE : epoch**0.65 ) # Make a copy of `model` if sched: _lowercase , _lowercase , _lowercase , _lowercase : Union[str, Any] = accelerator.prepare(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) else: _lowercase , _lowercase : Any = accelerator.prepare(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if sched: return (model, opt, sched, dataloader, ddp_model, ddp_opt, ddp_sched) return model, ddp_model, dataloader def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: # Test when on a single CPU or GPU that the context manager does nothing _lowercase , _lowercase , _lowercase : List[str] = get_training_setup(SCREAMING_SNAKE_CASE ) # Use a single batch _lowercase , _lowercase : List[Any] = next(iter(SCREAMING_SNAKE_CASE ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model _lowercase , _lowercase : List[Any] = accelerator.gather((ddp_input, ddp_target) ) _lowercase , _lowercase : Dict = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(SCREAMING_SNAKE_CASE ): step_model(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) else: # Sync grads step_model(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Since `no_sync` is a noop, `ddp_model` and `model` grads should always be in sync check_model_parameters(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue assert torch.allclose( param.grad , ddp_param.grad ), F"""Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})""" # Shuffle ddp_input on each iteration torch.manual_seed(1_337 + iteration ) _lowercase : List[Any] = ddp_input[torch.randperm(len(SCREAMING_SNAKE_CASE ) )] def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> Tuple: # Test on distributed setup that context manager behaves properly _lowercase , _lowercase , _lowercase : Dict = get_training_setup(SCREAMING_SNAKE_CASE ) # Use a single batch _lowercase , _lowercase : str = next(iter(SCREAMING_SNAKE_CASE ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model _lowercase , _lowercase : str = accelerator.gather((ddp_input, ddp_target) ) _lowercase , _lowercase : List[Any] = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(SCREAMING_SNAKE_CASE ): step_model(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) else: # Sync grads step_model(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue if iteration % 2 == 0: # Grads should not be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is False ), F"""Gradients in sync when they should not be:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})""" else: # Grads should be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is True ), F"""Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})""" # Shuffle ddp_input on each iteration torch.manual_seed(1_337 + iteration ) _lowercase : int = ddp_input[torch.randperm(len(SCREAMING_SNAKE_CASE ) )] def __magic_name__ ( SCREAMING_SNAKE_CASE=False , SCREAMING_SNAKE_CASE=False ) -> List[str]: _lowercase : List[str] = Accelerator( split_batches=SCREAMING_SNAKE_CASE , dispatch_batches=SCREAMING_SNAKE_CASE , gradient_accumulation_steps=2 ) # Test that context manager behaves properly _lowercase , _lowercase , _lowercase : Any = get_training_setup(SCREAMING_SNAKE_CASE ) for iteration, batch in enumerate(SCREAMING_SNAKE_CASE ): _lowercase , _lowercase : int = batch.values() # Gather the distributed inputs and targs for the base model _lowercase , _lowercase : Tuple = accelerator.gather((ddp_input, ddp_target) ) _lowercase , _lowercase : List[str] = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Do "gradient accumulation" (noop) with accelerator.accumulate(SCREAMING_SNAKE_CASE ): step_model(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue if ((iteration + 1) % 2 == 0) or (iteration == len(SCREAMING_SNAKE_CASE ) - 1): # Grads should be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is True ), F"""Gradients not in sync when they should be at iteration {iteration}:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})""" else: # Grads should not be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is False ), F"""Gradients in sync when they should not be at iteration {iteration}:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})""" # Shuffle ddp_input on each iteration torch.manual_seed(1_337 + iteration ) _lowercase : Optional[int] = ddp_input[torch.randperm(len(SCREAMING_SNAKE_CASE ) )] GradientState._reset_state() def __magic_name__ ( SCREAMING_SNAKE_CASE=False , SCREAMING_SNAKE_CASE=False ) -> Union[str, Any]: _lowercase : Dict = Accelerator( split_batches=SCREAMING_SNAKE_CASE , dispatch_batches=SCREAMING_SNAKE_CASE , gradient_accumulation_steps=2 ) # Test that context manager behaves properly _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase : int = get_training_setup(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for iteration, batch in enumerate(SCREAMING_SNAKE_CASE ): _lowercase , _lowercase : str = batch.values() # Gather the distributed inputs and targs for the base model _lowercase , _lowercase : Dict = accelerator.gather((ddp_input, ddp_target) ) _lowercase , _lowercase : Tuple = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" model.train() ddp_model.train() step_model(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) opt.step() if ((iteration + 1) % 2 == 0) or ((iteration + 1) == len(SCREAMING_SNAKE_CASE )): if split_batches: sched.step() else: for _ in range(accelerator.num_processes ): sched.step() opt.zero_grad() # Perform gradient accumulation under wrapper with accelerator.accumulate(SCREAMING_SNAKE_CASE ): step_model(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ddp_opt.step() ddp_sched.step() ddp_opt.zero_grad() # Learning rates should be the same assert ( opt.param_groups[0]["lr"] == ddp_opt.param_groups[0]["lr"] ), F"""Learning rates found in each optimizer did not align\nopt: {opt.param_groups[0]['lr']}\nDDP opt: {ddp_opt.param_groups[0]['lr']}\n""" _lowercase : Optional[int] = (((iteration + 1) % 2) == 0) or ((iteration + 1) == len(SCREAMING_SNAKE_CASE )) if accelerator.num_processes > 1: check_model_parameters(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Shuffle ddp_input on each iteration torch.manual_seed(1_337 + iteration ) GradientState._reset_state() def __magic_name__ ( ) -> Any: _lowercase : Any = Accelerator() _lowercase : List[Any] = RegressionDataset(length=80 ) _lowercase : List[str] = DataLoader(SCREAMING_SNAKE_CASE , batch_size=16 ) _lowercase : List[str] = RegressionDataset(length=96 ) _lowercase : Union[str, Any] = DataLoader(SCREAMING_SNAKE_CASE , batch_size=16 ) _lowercase , _lowercase : Tuple = accelerator.prepare(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) assert accelerator.gradient_state.active_dataloader is None for iteration, _ in enumerate(SCREAMING_SNAKE_CASE ): assert id(accelerator.gradient_state.active_dataloader ) == id(SCREAMING_SNAKE_CASE ) if iteration < len(SCREAMING_SNAKE_CASE ) - 1: assert not accelerator.gradient_state.end_of_dataloader if iteration == 1: for batch_num, _ in enumerate(SCREAMING_SNAKE_CASE ): assert id(accelerator.gradient_state.active_dataloader ) == id(SCREAMING_SNAKE_CASE ) if batch_num < len(SCREAMING_SNAKE_CASE ) - 1: assert not accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader assert accelerator.gradient_state.active_dataloader is None def __magic_name__ ( ) -> Dict: _lowercase : List[str] = Accelerator() _lowercase : List[Any] = accelerator.state if state.local_process_index == 0: print('**Test `accumulate` gradient accumulation with dataloader break**' ) test_dataloader_break() if state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print('**Test NOOP `no_sync` context manager**' ) test_noop_sync(SCREAMING_SNAKE_CASE ) if state.distributed_type in (DistributedType.MULTI_GPU, DistributedType.MULTI_CPU): if state.local_process_index == 0: print('**Test Distributed `no_sync` context manager**' ) test_distributed_sync(SCREAMING_SNAKE_CASE ) if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if state.local_process_index == 0: print( '**Test `accumulate` gradient accumulation, ' , F"""`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**""" , ) test_gradient_accumulation(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Currently will break on torch 2.0 +, need to investigate why if is_torch_version('<' , '2.0' ) or state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print( '**Test `accumulate` gradient accumulation with optimizer and scheduler, ' , '`split_batches=False`, `dispatch_batches=False`**' , ) test_gradient_accumulation_with_opt_and_scheduler() if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if not split_batch and not dispatch_batches: continue if state.local_process_index == 0: print( '**Test `accumulate` gradient accumulation with optimizer and scheduler, ' , F"""`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**""" , ) test_gradient_accumulation_with_opt_and_scheduler(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> Union[str, Any]: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
66
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: if a < 0 or b < 0: raise ValueError('the value of both inputs must be positive' ) _lowercase : str = str(bin(SCREAMING_SNAKE_CASE ) )[2:] # remove the leading "0b" _lowercase : int = str(bin(SCREAMING_SNAKE_CASE ) )[2:] # remove the leading "0b" _lowercase : List[Any] = max(len(SCREAMING_SNAKE_CASE ) , len(SCREAMING_SNAKE_CASE ) ) return "0b" + "".join( str(int(char_a == '1' and char_b == '1' ) ) for char_a, char_b in zip(a_binary.zfill(SCREAMING_SNAKE_CASE ) , b_binary.zfill(SCREAMING_SNAKE_CASE ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
66
1
import argparse import torch from torch import nn from transformers import MaMaaaConfig, MaMaaaForConditionalGeneration def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> Union[str, Any]: _lowercase : Optional[int] = [ 'encoder.version', 'decoder.version', 'model.encoder.version', 'model.decoder.version', 'decoder.output_projection.weight', '_float_tensor', 'encoder.embed_positions._float_tensor', 'decoder.embed_positions._float_tensor', ] for k in ignore_keys: state_dict.pop(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> Union[str, Any]: _lowercase , _lowercase : List[str] = emb.weight.shape _lowercase : str = nn.Linear(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , bias=SCREAMING_SNAKE_CASE ) _lowercase : Tuple = emb.weight.data return lin_layer def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> Any: _lowercase : List[Any] = torch.load(SCREAMING_SNAKE_CASE , map_location='cpu' ) _lowercase : Union[str, Any] = mam_aaa['args'] or mam_aaa['cfg']['model'] _lowercase : Optional[Any] = mam_aaa['model'] remove_ignore_keys_(SCREAMING_SNAKE_CASE ) _lowercase : int = state_dict['encoder.embed_tokens.weight'].shape[0] _lowercase : List[Any] = MaMaaaConfig( vocab_size=SCREAMING_SNAKE_CASE , max_position_embeddings=1_024 , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , encoder_layerdrop=args.encoder_layerdrop , decoder_layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='relu' , ) _lowercase : Union[str, Any] = state_dict['decoder.embed_tokens.weight'] _lowercase : Optional[int] = MaMaaaForConditionalGeneration(SCREAMING_SNAKE_CASE ) model.model.load_state_dict(SCREAMING_SNAKE_CASE , strict=SCREAMING_SNAKE_CASE ) _lowercase : List[str] = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument("fairseq_path", type=str, help="path to a model.pt on local filesystem.") parser.add_argument("pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") UpperCamelCase = parser.parse_args() UpperCamelCase = convert_fairseq_mamaaa_checkpoint_from_disk(args.fairseq_pathß) model.save_pretrained(args.pytorch_dump_folder_path)
66
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class lowerCAmelCase_ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCamelCase : int = IFInpaintingSuperResolutionPipeline _UpperCamelCase : Any = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"width", "height"} _UpperCamelCase : Union[str, Any] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"original_image"} ) _UpperCamelCase : Tuple = PipelineTesterMixin.required_optional_params - {"latents"} def __a ( self ): return self._get_superresolution_dummy_components() def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : int = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Union[str, Any] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : Union[str, Any] = floats_tensor((1, 3, 1_6, 1_6) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : Optional[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'original_image': original_image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def __a ( self ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def __a ( self ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != 'cuda' , reason='float16 requires CUDA' ) def __a ( self ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __a ( self ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __a ( self ): self._test_save_load_local() def __a ( self ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
66
1
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: if a < 0 or b < 0: raise ValueError('the value of both inputs must be positive' ) _lowercase : str = str(bin(SCREAMING_SNAKE_CASE ) )[2:] # remove the leading "0b" _lowercase : int = str(bin(SCREAMING_SNAKE_CASE ) )[2:] # remove the leading "0b" _lowercase : List[Any] = max(len(SCREAMING_SNAKE_CASE ) , len(SCREAMING_SNAKE_CASE ) ) return "0b" + "".join( str(int(char_a == '1' and char_b == '1' ) ) for char_a, char_b in zip(a_binary.zfill(SCREAMING_SNAKE_CASE ) , b_binary.zfill(SCREAMING_SNAKE_CASE ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
66
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> tuple[float, float]: # Check if the input is valid if not len(SCREAMING_SNAKE_CASE ) == len(SCREAMING_SNAKE_CASE ) == 3: raise ValueError('Please enter a valid equation.' ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError('Both a & b of two equations can\'t be zero.' ) # Extract the coefficients _lowercase , _lowercase , _lowercase : Tuple = equationa _lowercase , _lowercase , _lowercase : Dict = equationa # Calculate the determinants of the matrices _lowercase : str = aa * ba - aa * ba _lowercase : Any = ca * ba - ca * ba _lowercase : Optional[int] = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError('Infinite solutions. (Consistent system)' ) else: raise ValueError('No solution. (Inconsistent system)' ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: _lowercase : Union[str, Any] = determinant_x / determinant _lowercase : Tuple = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
66
1
from typing import Optional from .. import Features, NamedSplit from ..packaged_modules.text.text import Text from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class lowerCAmelCase_ ( __snake_case ): def __init__( self , _lowerCAmelCase , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = False , _lowerCAmelCase = False , _lowerCAmelCase = None , **_lowerCAmelCase , ): super().__init__( _lowerCAmelCase , split=_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase , keep_in_memory=_lowerCAmelCase , streaming=_lowerCAmelCase , num_proc=_lowerCAmelCase , **_lowerCAmelCase , ) _lowercase : Dict = path_or_paths if isinstance(_lowerCAmelCase , _lowerCAmelCase ) else {self.split: path_or_paths} _lowercase : Tuple = Text( cache_dir=_lowerCAmelCase , data_files=_lowerCAmelCase , features=_lowerCAmelCase , **_lowerCAmelCase , ) def __a ( self ): # Build iterable dataset if self.streaming: _lowercase : List[Any] = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: _lowercase : Any = None _lowercase : Optional[int] = None _lowercase : Union[str, Any] = None _lowercase : str = None self.builder.download_and_prepare( download_config=_lowerCAmelCase , download_mode=_lowerCAmelCase , verification_mode=_lowerCAmelCase , base_path=_lowerCAmelCase , num_proc=self.num_proc , ) _lowercase : Dict = self.builder.as_dataset( split=self.split , verification_mode=_lowerCAmelCase , in_memory=self.keep_in_memory ) return dataset
66
def __magic_name__ ( SCREAMING_SNAKE_CASE = 50 ) -> int: _lowercase : Optional[int] = [[0] * 3 for _ in range(length + 1 )] for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): different_colour_ways_number[row_length][tile_length - 2] += ( different_colour_ways_number[row_length - tile_start - tile_length][ tile_length - 2 ] + 1 ) return sum(different_colour_ways_number[length] ) if __name__ == "__main__": print(f'''{solution() = }''')
66
1
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import datasets import numpy as np import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, EvalPrediction, HfArgumentParser, PreTrainedTokenizer, TFAutoModelForSequenceClassification, TFTrainer, TFTrainingArguments, ) from transformers.utils import logging as hf_logging hf_logging.set_verbosity_info() hf_logging.enable_default_handler() hf_logging.enable_explicit_format() def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None , ) -> Tuple: _lowercase : List[str] = {} if train_file is not None: _lowercase : Tuple = [train_file] if eval_file is not None: _lowercase : Optional[Any] = [eval_file] if test_file is not None: _lowercase : Dict = [test_file] _lowercase : Dict = datasets.load_dataset('csv' , data_files=SCREAMING_SNAKE_CASE ) _lowercase : Dict = list(ds[list(files.keys() )[0]].features.keys() ) _lowercase : List[Any] = features_name.pop(SCREAMING_SNAKE_CASE ) _lowercase : Any = list(set(ds[list(files.keys() )[0]][label_name] ) ) _lowercase : Dict = {label: i for i, label in enumerate(SCREAMING_SNAKE_CASE )} _lowercase : Tuple = tokenizer.model_input_names _lowercase : int = {} if len(SCREAMING_SNAKE_CASE ) == 1: for k in files.keys(): _lowercase : List[str] = ds[k].map( lambda SCREAMING_SNAKE_CASE : tokenizer.batch_encode_plus( example[features_name[0]] , truncation=SCREAMING_SNAKE_CASE , max_length=SCREAMING_SNAKE_CASE , padding='max_length' ) , batched=SCREAMING_SNAKE_CASE , ) elif len(SCREAMING_SNAKE_CASE ) == 2: for k in files.keys(): _lowercase : Tuple = ds[k].map( lambda SCREAMING_SNAKE_CASE : tokenizer.batch_encode_plus( (example[features_name[0]], example[features_name[1]]) , truncation=SCREAMING_SNAKE_CASE , max_length=SCREAMING_SNAKE_CASE , padding='max_length' , ) , batched=SCREAMING_SNAKE_CASE , ) def gen_train(): for ex in transformed_ds[datasets.Split.TRAIN]: _lowercase : str = {k: v for k, v in ex.items() if k in input_names} _lowercase : List[str] = labelaid[ex[label_name]] yield (d, label) def gen_val(): for ex in transformed_ds[datasets.Split.VALIDATION]: _lowercase : str = {k: v for k, v in ex.items() if k in input_names} _lowercase : Tuple = labelaid[ex[label_name]] yield (d, label) def gen_test(): for ex in transformed_ds[datasets.Split.TEST]: _lowercase : int = {k: v for k, v in ex.items() if k in input_names} _lowercase : str = labelaid[ex[label_name]] yield (d, label) _lowercase : str = ( tf.data.Dataset.from_generator( SCREAMING_SNAKE_CASE , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.TRAIN in transformed_ds else None ) if train_ds is not None: _lowercase : Optional[int] = train_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TRAIN] ) ) ) _lowercase : int = ( tf.data.Dataset.from_generator( SCREAMING_SNAKE_CASE , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.VALIDATION in transformed_ds else None ) if val_ds is not None: _lowercase : str = val_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.VALIDATION] ) ) ) _lowercase : List[str] = ( tf.data.Dataset.from_generator( SCREAMING_SNAKE_CASE , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.TEST in transformed_ds else None ) if test_ds is not None: _lowercase : Optional[Any] = test_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TEST] ) ) ) return train_ds, val_ds, test_ds, labelaid UpperCamelCase = logging.getLogger(__name__) @dataclass class lowerCAmelCase_ : _UpperCamelCase : int = field(metadata={"help": "Which column contains the label"} ) _UpperCamelCase : str = field(default=__snake_case , metadata={"help": "The path of the training file"} ) _UpperCamelCase : Optional[str] = field(default=__snake_case , metadata={"help": "The path of the development file"} ) _UpperCamelCase : Optional[str] = field(default=__snake_case , metadata={"help": "The path of the test file"} ) _UpperCamelCase : int = field( default=128 , metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) _UpperCamelCase : bool = field( default=__snake_case , metadata={"help": "Overwrite the cached training and evaluation sets"} ) @dataclass class lowerCAmelCase_ : _UpperCamelCase : str = field( metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} ) _UpperCamelCase : Optional[str] = field( default=__snake_case , metadata={"help": "Pretrained config name or path if not the same as model_name"} ) _UpperCamelCase : Optional[str] = field( default=__snake_case , metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"} ) _UpperCamelCase : bool = field(default=__snake_case , metadata={"help": "Set this flag to use fast tokenization."} ) # If you want to tweak more attributes on your tokenizer, you should do it in a distinct script, # or just modify its tokenizer_config.json. _UpperCamelCase : Optional[str] = field( default=__snake_case , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) def __magic_name__ ( ) -> str: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _lowercase : Optional[int] = HfArgumentParser((ModelArguments, DataTrainingArguments, TFTrainingArguments) ) _lowercase , _lowercase , _lowercase : Optional[int] = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" ' --overwrite_output_dir to overcome.' ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , level=logging.INFO , ) logger.info( F"""n_replicas: {training_args.n_replicas}, distributed training: {bool(training_args.n_replicas > 1 )}, """ F"""16-bits training: {training_args.fpaa}""" ) logger.info(F"""Training/evaluation parameters {training_args}""" ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _lowercase : List[str] = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) _lowercase , _lowercase , _lowercase , _lowercase : int = get_tfds( train_file=data_args.train_file , eval_file=data_args.dev_file , test_file=data_args.test_file , tokenizer=SCREAMING_SNAKE_CASE , label_column_id=data_args.label_column_id , max_seq_length=data_args.max_seq_length , ) _lowercase : List[Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=len(SCREAMING_SNAKE_CASE ) , labelaid=SCREAMING_SNAKE_CASE , idalabel={id: label for label, id in labelaid.items()} , finetuning_task='text-classification' , cache_dir=model_args.cache_dir , ) with training_args.strategy.scope(): _lowercase : List[str] = TFAutoModelForSequenceClassification.from_pretrained( model_args.model_name_or_path , from_pt=bool('.bin' in model_args.model_name_or_path ) , config=SCREAMING_SNAKE_CASE , cache_dir=model_args.cache_dir , ) def compute_metrics(SCREAMING_SNAKE_CASE ) -> Dict: _lowercase : Optional[int] = np.argmax(p.predictions , axis=1 ) return {"acc": (preds == p.label_ids).mean()} # Initialize our Trainer _lowercase : Optional[Any] = TFTrainer( model=SCREAMING_SNAKE_CASE , args=SCREAMING_SNAKE_CASE , train_dataset=SCREAMING_SNAKE_CASE , eval_dataset=SCREAMING_SNAKE_CASE , compute_metrics=SCREAMING_SNAKE_CASE , ) # Training if training_args.do_train: trainer.train() trainer.save_model() tokenizer.save_pretrained(training_args.output_dir ) # Evaluation _lowercase : Dict = {} if training_args.do_eval: logger.info('*** Evaluate ***' ) _lowercase : int = trainer.evaluate() _lowercase : List[Any] = os.path.join(training_args.output_dir , 'eval_results.txt' ) with open(SCREAMING_SNAKE_CASE , 'w' ) as writer: logger.info('***** Eval results *****' ) for key, value in result.items(): logger.info(F""" {key} = {value}""" ) writer.write(F"""{key} = {value}\n""" ) results.update(SCREAMING_SNAKE_CASE ) return results if __name__ == "__main__": main()
66
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) UpperCamelCase = { "configuration_convnext": ["CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP", "ConvNextConfig", "ConvNextOnnxConfig"] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["ConvNextFeatureExtractor"] UpperCamelCase = ["ConvNextImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST", "ConvNextForImageClassification", "ConvNextModel", "ConvNextPreTrainedModel", "ConvNextBackbone", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "TFConvNextForImageClassification", "TFConvNextModel", "TFConvNextPreTrainedModel", ] if TYPE_CHECKING: from .configuration_convnext import CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvNextConfig, ConvNextOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_convnext import ConvNextFeatureExtractor from .image_processing_convnext import ConvNextImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convnext import ( CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvNextBackbone, ConvNextForImageClassification, ConvNextModel, ConvNextPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convnext import TFConvNextForImageClassification, TFConvNextModel, TFConvNextPreTrainedModel else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure)
66
1
import argparse import hashlib # hashlib is only used inside the Test class import struct class lowerCAmelCase_ : def __init__( self , _lowerCAmelCase ): _lowercase : int = data _lowercase : Optional[int] = [0x67_452_301, 0xef_cda_b89, 0x98_bad_cfe, 0x10_325_476, 0xc3_d2e_1f0] @staticmethod def __a ( _lowerCAmelCase , _lowerCAmelCase ): return ((n << b) | (n >> (3_2 - b))) & 0xff_fff_fff def __a ( self ): _lowercase : Union[str, Any] = b'\x80' + b'\x00' * (6_3 - (len(self.data ) + 8) % 6_4) _lowercase : List[Any] = self.data + padding + struct.pack('>Q' , 8 * len(self.data ) ) return padded_data def __a ( self ): return [ self.padded_data[i : i + 6_4] for i in range(0 , len(self.padded_data ) , 6_4 ) ] def __a ( self , _lowerCAmelCase ): _lowercase : str = list(struct.unpack('>16L' , _lowerCAmelCase ) ) + [0] * 6_4 for i in range(1_6 , 8_0 ): _lowercase : List[Any] = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 1_4] ^ w[i - 1_6]) , 1 ) return w def __a ( self ): _lowercase : Dict = self.padding() _lowercase : int = self.split_blocks() for block in self.blocks: _lowercase : List[str] = self.expand_block(_lowerCAmelCase ) _lowercase , _lowercase , _lowercase , _lowercase , _lowercase : List[Any] = self.h for i in range(0 , 8_0 ): if 0 <= i < 2_0: _lowercase : Any = (b & c) | ((~b) & d) _lowercase : List[Any] = 0x5a_827_999 elif 2_0 <= i < 4_0: _lowercase : Optional[Any] = b ^ c ^ d _lowercase : Union[str, Any] = 0x6e_d9e_ba1 elif 4_0 <= i < 6_0: _lowercase : Optional[int] = (b & c) | (b & d) | (c & d) _lowercase : Tuple = 0x8f_1bb_cdc elif 6_0 <= i < 8_0: _lowercase : Any = b ^ c ^ d _lowercase : Optional[Any] = 0xca_62c_1d6 _lowercase , _lowercase , _lowercase , _lowercase , _lowercase : Union[str, Any] = ( self.rotate(_lowerCAmelCase , 5 ) + f + e + k + expanded_block[i] & 0xff_fff_fff, a, self.rotate(_lowerCAmelCase , 3_0 ), c, d, ) _lowercase : Optional[Any] = ( self.h[0] + a & 0xff_fff_fff, self.h[1] + b & 0xff_fff_fff, self.h[2] + c & 0xff_fff_fff, self.h[3] + d & 0xff_fff_fff, self.h[4] + e & 0xff_fff_fff, ) return ("{:08x}" * 5).format(*self.h ) def __magic_name__ ( ) -> List[Any]: _lowercase : Union[str, Any] = b'Test String' assert SHAaHash(SCREAMING_SNAKE_CASE ).final_hash() == hashlib.shaa(SCREAMING_SNAKE_CASE ).hexdigest() # noqa: S324 def __magic_name__ ( ) -> List[str]: _lowercase : List[Any] = argparse.ArgumentParser(description='Process some strings or files' ) parser.add_argument( '--string' , dest='input_string' , default='Hello World!! Welcome to Cryptography' , help='Hash the string' , ) parser.add_argument('--file' , dest='input_file' , help='Hash contents of a file' ) _lowercase : Optional[Any] = parser.parse_args() _lowercase : Any = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file , 'rb' ) as f: _lowercase : Dict = f.read() else: _lowercase : List[Any] = bytes(SCREAMING_SNAKE_CASE , 'utf-8' ) print(SHAaHash(SCREAMING_SNAKE_CASE ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
66
def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> bool: if p < 2: raise ValueError('p should not be less than 2!' ) elif p == 2: return True _lowercase : Optional[Any] = 4 _lowercase : Tuple = (1 << p) - 1 for _ in range(p - 2 ): _lowercase : Union[str, Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
66
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = "▁" UpperCamelCase = {"vocab_file": "sentencepiece.bpe.model", "monolingual_vocab_file": "dict.txt"} UpperCamelCase = { "vocab_file": { "vinai/bartpho-syllable": "https://huggingface.co/vinai/bartpho-syllable/resolve/main/sentencepiece.bpe.model", }, "monolingual_vocab_file": { "vinai/bartpho-syllable": "https://huggingface.co/vinai/bartpho-syllable/resolve/main/dict.txt", }, } UpperCamelCase = {"vinai/bartpho-syllable": 1_024} class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : List[str] = VOCAB_FILES_NAMES _UpperCamelCase : Optional[int] = PRETRAINED_VOCAB_FILES_MAP _UpperCamelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCamelCase : Tuple = ["input_ids", "attention_mask"] def __init__( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase="<s>" , _lowerCAmelCase="</s>" , _lowerCAmelCase="</s>" , _lowerCAmelCase="<s>" , _lowerCAmelCase="<unk>" , _lowerCAmelCase="<pad>" , _lowerCAmelCase="<mask>" , _lowerCAmelCase = None , **_lowerCAmelCase , ): # Mask token behave like a normal word, i.e. include the space before it _lowercase : str = AddedToken(_lowerCAmelCase , lstrip=_lowerCAmelCase , rstrip=_lowerCAmelCase ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ) else mask_token _lowercase : Optional[Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=_lowerCAmelCase , eos_token=_lowerCAmelCase , unk_token=_lowerCAmelCase , sep_token=_lowerCAmelCase , cls_token=_lowerCAmelCase , pad_token=_lowerCAmelCase , mask_token=_lowerCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **_lowerCAmelCase , ) _lowercase : Optional[Any] = vocab_file _lowercase : Dict = monolingual_vocab_file _lowercase : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(_lowerCAmelCase ) ) # Load the reduced vocab # Keep order of special tokens for backward compatibility _lowercase : List[Any] = {} _lowercase : Union[str, Any] = 0 for token in [bos_token, pad_token, eos_token, unk_token, sep_token, cls_token]: if str(_lowerCAmelCase ) not in self.fairseq_tokens_to_ids: _lowercase : Union[str, Any] = cnt cnt += 1 with open(_lowerCAmelCase , 'r' , encoding='utf-8' ) as f: for line in f.readlines(): _lowercase : str = line.strip().split()[0] _lowercase : Optional[Any] = len(self.fairseq_tokens_to_ids ) if str(_lowerCAmelCase ) not in self.fairseq_tokens_to_ids: _lowercase : List[str] = len(self.fairseq_tokens_to_ids ) _lowercase : Optional[int] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self ): _lowercase : Dict = self.__dict__.copy() _lowercase : Any = None _lowercase : Optional[int] = self.sp_model.serialized_model_proto() return state def __setstate__( self , _lowerCAmelCase ): _lowercase : str = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): _lowercase : str = {} _lowercase : str = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _lowercase : int = [self.cls_token_id] _lowercase : Optional[int] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __a ( self , _lowerCAmelCase , _lowerCAmelCase = None , _lowerCAmelCase = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_lowerCAmelCase , token_ids_a=_lowerCAmelCase , already_has_special_tokens=_lowerCAmelCase ) if token_ids_a is None: return [1] + ([0] * len(_lowerCAmelCase )) + [1] return [1] + ([0] * len(_lowerCAmelCase )) + [1, 1] + ([0] * len(_lowerCAmelCase )) + [1] def __a ( self , _lowerCAmelCase , _lowerCAmelCase = None ): _lowercase : List[Any] = [self.sep_token_id] _lowercase : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __a ( self ): return len(self.fairseq_ids_to_tokens ) def __a ( self ): _lowercase : Any = {self.convert_ids_to_tokens(_lowerCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __a ( self , _lowerCAmelCase ): return self.sp_model.encode(_lowerCAmelCase , out_type=_lowerCAmelCase ) def __a ( self , _lowerCAmelCase ): if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] else: return self.unk_token_id def __a ( self , _lowerCAmelCase ): return self.fairseq_ids_to_tokens[index] def __a ( self , _lowerCAmelCase ): _lowercase : Tuple = ''.join(_lowerCAmelCase ).replace(_lowerCAmelCase , ' ' ).strip() return out_string def __a ( self , _lowerCAmelCase , _lowerCAmelCase = None ): if not os.path.isdir(_lowerCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _lowercase : Dict = os.path.join( _lowerCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) _lowercase : Union[str, Any] = os.path.join( _lowerCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['monolingual_vocab_file'] , ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_lowerCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _lowerCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(_lowerCAmelCase , 'wb' ) as fi: _lowercase : Optional[int] = self.sp_model.serialized_model_proto() fi.write(_lowerCAmelCase ) if os.path.abspath(self.monolingual_vocab_file ) != os.path.abspath( _lowerCAmelCase ) and os.path.isfile(self.monolingual_vocab_file ): copyfile(self.monolingual_vocab_file , _lowerCAmelCase ) elif not os.path.isfile(self.monolingual_vocab_file ): with open(_lowerCAmelCase , 'w' , encoding='utf-8' ) as fp: for token in self.fairseq_tokens_to_ids: if token not in self.all_special_tokens: fp.write(F"""{str(_lowerCAmelCase )} \n""" ) return out_vocab_file, out_monolingual_vocab_file
66
import random def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = False ) -> dict: _lowercase : dict = {i: [] for i in range(SCREAMING_SNAKE_CASE )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(SCREAMING_SNAKE_CASE ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(SCREAMING_SNAKE_CASE ): for j in range(i + 1 , SCREAMING_SNAKE_CASE ): if random.random() < probability: graph[i].append(SCREAMING_SNAKE_CASE ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(SCREAMING_SNAKE_CASE ) return graph def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> dict: return { i: [j for j in range(SCREAMING_SNAKE_CASE ) if i != j] for i in range(SCREAMING_SNAKE_CASE ) } if __name__ == "__main__": import doctest doctest.testmod()
66
1
from collections import defaultdict def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> bool: _lowercase : Union[str, Any] = first_str.lower().strip() _lowercase : Tuple = second_str.lower().strip() # Remove whitespace _lowercase : Dict = first_str.replace(' ' , '' ) _lowercase : str = second_str.replace(' ' , '' ) # Strings of different lengths are not anagrams if len(SCREAMING_SNAKE_CASE ) != len(SCREAMING_SNAKE_CASE ): return False # Default values for count should be 0 _lowercase : defaultdict[str, int] = defaultdict(SCREAMING_SNAKE_CASE ) # For each character in input strings, # increment count in the corresponding for i in range(len(SCREAMING_SNAKE_CASE ) ): count[first_str[i]] += 1 count[second_str[i]] -= 1 return all(_count == 0 for _count in count.values() ) if __name__ == "__main__": from doctest import testmod testmod() UpperCamelCase = input("Enter the first string ").strip() UpperCamelCase = input("Enter the second string ").strip() UpperCamelCase = check_anagrams(input_a, input_b) print(f'''{input_a} and {input_b} are {"" if status else "not "}anagrams.''')
66
from __future__ import annotations UpperCamelCase = tuple[int, int, int] UpperCamelCase = tuple[str, str, str] # used alphabet -------------------------- # from string.ascii_uppercase UpperCamelCase = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" # -------------------------- default selection -------------------------- # rotors -------------------------- UpperCamelCase = "EGZWVONAHDCLFQMSIPJBYUKXTR" UpperCamelCase = "FOBHMDKEXQNRAULPGSJVTYICZW" UpperCamelCase = "ZJXESIUQLHAVRMDOYGTNFWPBKC" # reflector -------------------------- UpperCamelCase = { "A": "N", "N": "A", "B": "O", "O": "B", "C": "P", "P": "C", "D": "Q", "Q": "D", "E": "R", "R": "E", "F": "S", "S": "F", "G": "T", "T": "G", "H": "U", "U": "H", "I": "V", "V": "I", "J": "W", "W": "J", "K": "X", "X": "K", "L": "Y", "Y": "L", "M": "Z", "Z": "M", } # -------------------------- extra rotors -------------------------- UpperCamelCase = "RMDJXFUWGISLHVTCQNKYPBEZOA" UpperCamelCase = "SGLCPQWZHKXAREONTFBVIYJUDM" UpperCamelCase = "HVSICLTYKQUBXDWAJZOMFGPREN" UpperCamelCase = "RZWQHFMVDBKICJLNTUXAGYPSOE" UpperCamelCase = "LFKIJODBEGAMQPXVUHYSTCZRWN" UpperCamelCase = "KOAEGVDHXPQZMLFTYWJNBRCIUS" def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> tuple[RotorPositionT, RotorSelectionT, dict[str, str]]: # Checks if there are 3 unique rotors if (unique_rotsel := len(set(SCREAMING_SNAKE_CASE ) )) < 3: _lowercase : Optional[int] = F"""Please use 3 unique rotors (not {unique_rotsel})""" raise Exception(SCREAMING_SNAKE_CASE ) # Checks if rotor positions are valid _lowercase , _lowercase , _lowercase : int = rotpos if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : Dict = F"""First rotor position is not within range of 1..26 ({rotorposa}""" raise ValueError(SCREAMING_SNAKE_CASE ) if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : int = F"""Second rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(SCREAMING_SNAKE_CASE ) if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : str = F"""Third rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(SCREAMING_SNAKE_CASE ) # Validates string and returns dict _lowercase : Tuple = _plugboard(SCREAMING_SNAKE_CASE ) return rotpos, rotsel, pbdict def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> dict[str, str]: # tests the input string if it # a) is type string # b) has even length (so pairs can be made) if not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): _lowercase : Optional[int] = F"""Plugboard setting isn't type string ({type(SCREAMING_SNAKE_CASE )})""" raise TypeError(SCREAMING_SNAKE_CASE ) elif len(SCREAMING_SNAKE_CASE ) % 2 != 0: _lowercase : Optional[int] = F"""Odd number of symbols ({len(SCREAMING_SNAKE_CASE )})""" raise Exception(SCREAMING_SNAKE_CASE ) elif pbstring == "": return {} pbstring.replace(' ' , '' ) # Checks if all characters are unique _lowercase : Dict = set() for i in pbstring: if i not in abc: _lowercase : str = F"""'{i}' not in list of symbols""" raise Exception(SCREAMING_SNAKE_CASE ) elif i in tmppbl: _lowercase : int = F"""Duplicate symbol ({i})""" raise Exception(SCREAMING_SNAKE_CASE ) else: tmppbl.add(SCREAMING_SNAKE_CASE ) del tmppbl # Created the dictionary _lowercase : Optional[Any] = {} for j in range(0 , len(SCREAMING_SNAKE_CASE ) - 1 , 2 ): _lowercase : Dict = pbstring[j + 1] _lowercase : Union[str, Any] = pbstring[j] return pb def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = (rotora, rotora, rotora) , SCREAMING_SNAKE_CASE = "" , ) -> str: _lowercase : List[str] = text.upper() _lowercase , _lowercase , _lowercase : List[str] = _validator( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , plugb.upper() ) _lowercase , _lowercase , _lowercase : Optional[int] = rotor_position _lowercase , _lowercase , _lowercase : Union[str, Any] = rotor_selection rotorposa -= 1 rotorposa -= 1 rotorposa -= 1 _lowercase : Optional[int] = [] # encryption/decryption process -------------------------- for symbol in text: if symbol in abc: # 1st plugboard -------------------------- if symbol in plugboard: _lowercase : Dict = plugboard[symbol] # rotor ra -------------------------- _lowercase : Optional[Any] = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : Union[str, Any] = rotora[index % len(SCREAMING_SNAKE_CASE )] # rotor rb -------------------------- _lowercase : Tuple = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : str = rotora[index % len(SCREAMING_SNAKE_CASE )] # rotor rc -------------------------- _lowercase : List[Any] = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : List[str] = rotora[index % len(SCREAMING_SNAKE_CASE )] # reflector -------------------------- # this is the reason you don't need another machine to decipher _lowercase : List[str] = reflector[symbol] # 2nd rotors _lowercase : List[str] = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] _lowercase : Tuple = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] _lowercase : Dict = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] # 2nd plugboard if symbol in plugboard: _lowercase : int = plugboard[symbol] # moves/resets rotor positions rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : Any = 0 rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : int = 0 rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : Any = 0 # else: # pass # Error could be also raised # raise ValueError( # 'Invalid symbol('+repr(symbol)+')') result.append(SCREAMING_SNAKE_CASE ) return "".join(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": UpperCamelCase = "This is my Python script that emulates the Enigma machine from WWII." UpperCamelCase = (1, 1, 1) UpperCamelCase = "pictures" UpperCamelCase = (rotora, rotora, rotora) UpperCamelCase = enigma(message, rotor_pos, rotor_sel, pb) print("Encrypted message:", en) print("Decrypted message:", enigma(en, rotor_pos, rotor_sel, pb))
66
1
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionInstructPixaPixPipeline, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.utils import floats_tensor, load_image, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowerCAmelCase_ ( __snake_case , __snake_case , __snake_case , unittest.TestCase ): _UpperCamelCase : Dict = StableDiffusionInstructPixaPixPipeline _UpperCamelCase : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"height", "width", "cross_attention_kwargs"} _UpperCamelCase : Dict = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS _UpperCamelCase : List[Any] = IMAGE_TO_IMAGE_IMAGE_PARAMS _UpperCamelCase : Optional[Any] = IMAGE_TO_IMAGE_IMAGE_PARAMS def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[Any] = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=8 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=3_2 , ) _lowercase : Optional[int] = PNDMScheduler(skip_prk_steps=_lowerCAmelCase ) torch.manual_seed(0 ) _lowercase : List[Any] = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) _lowercase : int = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) _lowercase : Any = CLIPTextModel(_lowerCAmelCase ) _lowercase : Tuple = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) _lowercase : List[str] = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): _lowercase : Tuple = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : Optional[int] = image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowercase : Dict = Image.fromarray(np.uinta(_lowerCAmelCase ) ).convert('RGB' ) if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : List[Any] = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : int = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : Any = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'image_guidance_scale': 1, 'output_type': 'numpy', } return inputs def __a ( self ): _lowercase : str = 'cpu' # ensure determinism for the device-dependent torch.Generator _lowercase : List[Any] = self.get_dummy_components() _lowercase : Optional[int] = StableDiffusionInstructPixaPixPipeline(**_lowerCAmelCase ) _lowercase : List[str] = sd_pipe.to(_lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : List[Any] = self.get_dummy_inputs(_lowerCAmelCase ) _lowercase : Any = sd_pipe(**_lowerCAmelCase ).images _lowercase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) _lowercase : List[str] = np.array([0.75_26, 0.37_50, 0.45_47, 0.61_17, 0.58_66, 0.50_16, 0.43_27, 0.56_42, 0.48_15] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def __a ( self ): _lowercase : Any = 'cpu' # ensure determinism for the device-dependent torch.Generator _lowercase : int = self.get_dummy_components() _lowercase : Union[str, Any] = StableDiffusionInstructPixaPixPipeline(**_lowerCAmelCase ) _lowercase : List[Any] = sd_pipe.to(_lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Tuple = self.get_dummy_inputs(_lowerCAmelCase ) _lowercase : Optional[Any] = 'french fries' _lowercase : Dict = sd_pipe(**_lowerCAmelCase , negative_prompt=_lowerCAmelCase ) _lowercase : Optional[Any] = output.images _lowercase : Union[str, Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 3_2, 3_2, 3) _lowercase : Any = np.array([0.75_11, 0.36_42, 0.45_53, 0.62_36, 0.57_97, 0.50_13, 0.43_43, 0.56_11, 0.48_31] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def __a ( self ): _lowercase : Union[str, Any] = 'cpu' # ensure determinism for the device-dependent torch.Generator _lowercase : Dict = self.get_dummy_components() _lowercase : int = StableDiffusionInstructPixaPixPipeline(**_lowerCAmelCase ) _lowercase : Any = sd_pipe.to(_lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Any = self.get_dummy_inputs(_lowerCAmelCase ) _lowercase : Union[str, Any] = [inputs['prompt']] * 2 _lowercase : Union[str, Any] = np.array(inputs['image'] ).astype(np.floataa ) / 2_55.0 _lowercase : Tuple = torch.from_numpy(_lowerCAmelCase ).unsqueeze(0 ).to(_lowerCAmelCase ) _lowercase : Optional[int] = image / 2 + 0.5 _lowercase : List[Any] = image.permute(0 , 3 , 1 , 2 ) _lowercase : Optional[int] = image.repeat(2 , 1 , 1 , 1 ) _lowercase : Any = sd_pipe(**_lowerCAmelCase ).images _lowercase : List[str] = image[-1, -3:, -3:, -1] assert image.shape == (2, 3_2, 3_2, 3) _lowercase : Optional[int] = np.array([0.58_12, 0.57_48, 0.52_22, 0.59_08, 0.56_95, 0.71_74, 0.68_04, 0.55_23, 0.55_79] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def __a ( self ): _lowercase : Optional[Any] = 'cpu' # ensure determinism for the device-dependent torch.Generator _lowercase : Dict = self.get_dummy_components() _lowercase : Any = EulerAncestralDiscreteScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='scaled_linear' ) _lowercase : Optional[Any] = StableDiffusionInstructPixaPixPipeline(**_lowerCAmelCase ) _lowercase : List[str] = sd_pipe.to(_lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Dict = self.get_dummy_inputs(_lowerCAmelCase ) _lowercase : List[str] = sd_pipe(**_lowerCAmelCase ).images _lowercase : List[Any] = image[0, -3:, -3:, -1] _lowercase : Optional[int] = [round(_lowerCAmelCase , 4 ) for x in image_slice.flatten().tolist()] print(','.join([str(_lowerCAmelCase ) for x in slice] ) ) assert image.shape == (1, 3_2, 3_2, 3) _lowercase : str = np.array([0.74_17, 0.38_42, 0.47_32, 0.57_76, 0.58_91, 0.51_39, 0.40_52, 0.56_73, 0.49_86] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def __a ( self ): super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def __a ( self ): _lowercase : Union[str, Any] = self.get_dummy_components() _lowercase : Dict = StableDiffusionInstructPixaPixPipeline(**_lowerCAmelCase ) _lowercase : List[str] = VaeImageProcessor(do_resize=_lowerCAmelCase , do_normalize=_lowerCAmelCase ) _lowercase : Dict = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : str = pipe(**self.get_dummy_inputs_by_type(_lowerCAmelCase , input_image_type='pt' ) )[0] _lowercase : List[str] = components['vae'] _lowercase : Optional[Any] = self.get_dummy_inputs_by_type(_lowerCAmelCase , input_image_type='pt' ) for image_param in self.image_latents_params: if image_param in inputs.keys(): _lowercase : Optional[int] = vae.encode(inputs[image_param] ).latent_dist.mode() _lowercase : Optional[Any] = pipe(**_lowerCAmelCase )[0] _lowercase : List[str] = np.abs(out - out_latents_inputs ).max() self.assertLess(_lowerCAmelCase , 1E-4 , 'passing latents as image input generate different result from passing image' ) @slow @require_torch_gpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self , _lowerCAmelCase=0 ): _lowercase : Tuple = torch.manual_seed(_lowerCAmelCase ) _lowercase : List[str] = load_image( 'https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/stable_diffusion_pix2pix/example.jpg' ) _lowercase : Optional[int] = { 'prompt': 'turn him into a cyborg', 'image': image, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'image_guidance_scale': 1.0, 'output_type': 'numpy', } return inputs def __a ( self ): _lowercase : Optional[Any] = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=_lowerCAmelCase ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) pipe.enable_attention_slicing() _lowercase : Tuple = self.get_inputs() _lowercase : Dict = pipe(**_lowerCAmelCase ).images _lowercase : Any = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) _lowercase : Optional[Any] = np.array([0.59_02, 0.60_15, 0.60_27, 0.59_83, 0.60_92, 0.60_61, 0.57_65, 0.57_85, 0.55_55] ) assert np.abs(expected_slice - image_slice ).max() < 1E-3 def __a ( self ): _lowercase : str = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=_lowerCAmelCase ) _lowercase : Dict = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) pipe.enable_attention_slicing() _lowercase : Optional[int] = self.get_inputs() _lowercase : Optional[int] = pipe(**_lowerCAmelCase ).images _lowercase : Any = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) _lowercase : List[Any] = np.array([0.65_78, 0.68_17, 0.69_72, 0.67_61, 0.68_56, 0.69_16, 0.64_28, 0.65_16, 0.63_01] ) assert np.abs(expected_slice - image_slice ).max() < 1E-3 def __a ( self ): _lowercase : Tuple = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=_lowerCAmelCase ) _lowercase : Optional[int] = DDIMScheduler.from_config(pipe.scheduler.config ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) pipe.enable_attention_slicing() _lowercase : Tuple = self.get_inputs() _lowercase : int = pipe(**_lowerCAmelCase ).images _lowercase : Tuple = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_1_2, 5_1_2, 3) _lowercase : str = np.array([0.38_28, 0.38_34, 0.38_18, 0.37_92, 0.38_65, 0.37_52, 0.37_92, 0.38_47, 0.37_53] ) assert np.abs(expected_slice - image_slice ).max() < 1E-3 def __a ( self ): _lowercase : Dict = 0 def callback_fn(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> None: _lowercase : Any = True nonlocal number_of_steps number_of_steps += 1 if step == 1: _lowercase : Optional[int] = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) _lowercase : Dict = latents[0, -3:, -3:, -1] _lowercase : Any = np.array([-0.24_63, -0.46_44, -0.97_56, 1.51_76, 1.44_14, 0.78_66, 0.98_97, 0.85_21, 0.79_83] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5E-2 elif step == 2: _lowercase : List[str] = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 6_4, 6_4) _lowercase : List[Any] = latents[0, -3:, -3:, -1] _lowercase : str = np.array([-0.26_44, -0.46_26, -0.96_53, 1.51_76, 1.45_51, 0.76_86, 0.98_05, 0.84_52, 0.81_15] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5E-2 _lowercase : Tuple = False _lowercase : Union[str, Any] = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=_lowerCAmelCase , torch_dtype=torch.floataa ) _lowercase : str = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) pipe.enable_attention_slicing() _lowercase : Any = self.get_inputs() pipe(**_lowerCAmelCase , callback=_lowerCAmelCase , callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def __a ( self ): torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() _lowercase : Dict = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=_lowerCAmelCase , torch_dtype=torch.floataa ) _lowercase : Dict = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() _lowercase : List[Any] = self.get_inputs() _lowercase : List[Any] = pipe(**_lowerCAmelCase ) _lowercase : Optional[int] = torch.cuda.max_memory_allocated() # make sure that less than 2.2 GB is allocated assert mem_bytes < 2.2 * 1_0**9 def __a ( self ): _lowercase : Union[str, Any] = self.get_inputs() # resize to resolution that is divisible by 8 but not 16 or 32 _lowercase : Union[str, Any] = inputs['image'].resize((5_0_4, 5_0_4) ) _lowercase : List[str] = 'timbrooks/instruct-pix2pix' _lowercase : Dict = StableDiffusionInstructPixaPixPipeline.from_pretrained( _lowerCAmelCase , safety_checker=_lowerCAmelCase , ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) pipe.enable_attention_slicing() _lowercase : Any = pipe(**_lowerCAmelCase ) _lowercase : List[str] = output.images[0] _lowercase : List[Any] = image[2_5_5:2_5_8, 3_8_3:3_8_6, -1] assert image.shape == (5_0_4, 5_0_4, 3) _lowercase : Tuple = np.array([0.27_26, 0.25_29, 0.26_64, 0.26_55, 0.26_41, 0.26_42, 0.25_91, 0.26_49, 0.25_90] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-3
66
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = {"configuration_glpn": ["GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP", "GLPNConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["GLPNFeatureExtractor"] UpperCamelCase = ["GLPNImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "GLPN_PRETRAINED_MODEL_ARCHIVE_LIST", "GLPNForDepthEstimation", "GLPNLayer", "GLPNModel", "GLPNPreTrainedModel", ] if TYPE_CHECKING: from .configuration_glpn import GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP, GLPNConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_glpn import GLPNFeatureExtractor from .image_processing_glpn import GLPNImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_glpn import ( GLPN_PRETRAINED_MODEL_ARCHIVE_LIST, GLPNForDepthEstimation, GLPNLayer, GLPNModel, GLPNPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
1
def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: if divisor % 5 == 0 or divisor % 2 == 0: return 0 _lowercase : int = 1 _lowercase : List[str] = 1 while repunit: _lowercase : Union[str, Any] = (10 * repunit + 1) % divisor repunit_index += 1 return repunit_index def __magic_name__ ( SCREAMING_SNAKE_CASE = 1_000_000 ) -> int: _lowercase : Tuple = limit - 1 if divisor % 2 == 0: divisor += 1 while least_divisible_repunit(SCREAMING_SNAKE_CASE ) <= limit: divisor += 2 return divisor if __name__ == "__main__": print(f'''{solution() = }''')
66
import unittest from transformers import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device if is_torch_available(): import torch from transformers import AutoModelForImageClassification if is_vision_available(): from transformers import AutoImageProcessor @require_torch @require_vision class lowerCAmelCase_ ( unittest.TestCase ): @slow def __a ( self ): _lowercase : List[Any] = AutoImageProcessor.from_pretrained('microsoft/dit-base-finetuned-rvlcdip' ) _lowercase : List[Any] = AutoModelForImageClassification.from_pretrained('microsoft/dit-base-finetuned-rvlcdip' ) model.to(_lowerCAmelCase ) from datasets import load_dataset _lowercase : Union[str, Any] = load_dataset('nielsr/rvlcdip-demo' ) _lowercase : Any = dataset['train'][0]['image'].convert('RGB' ) _lowercase : List[str] = image_processor(_lowerCAmelCase , return_tensors='pt' ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): _lowercase : Dict = model(**_lowerCAmelCase ) _lowercase : Any = outputs.logits _lowercase : str = torch.Size((1, 1_6) ) self.assertEqual(logits.shape , _lowerCAmelCase ) _lowercase : Union[str, Any] = torch.tensor( [-0.41_58, -0.40_92, -0.43_47] , device=_lowerCAmelCase , dtype=torch.float , ) self.assertTrue(torch.allclose(logits[0, :3] , _lowerCAmelCase , atol=1E-4 ) )
66
1
import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class lowerCAmelCase_ : def __init__( self , _lowerCAmelCase , _lowerCAmelCase=1_3 , _lowerCAmelCase=2 , _lowerCAmelCase=2_4 , _lowerCAmelCase=1_6 , _lowerCAmelCase=True , _lowerCAmelCase=True , _lowerCAmelCase=3_2 , _lowerCAmelCase=5 , _lowerCAmelCase=4 , _lowerCAmelCase=3_7 , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.1 , _lowerCAmelCase=1_0 , _lowerCAmelCase=0.02 , _lowerCAmelCase=None , _lowerCAmelCase=2 , _lowerCAmelCase=2 , ): _lowercase : List[Any] = parent _lowercase : Dict = batch_size _lowercase : int = patch_size _lowercase : int = max_length _lowercase : Optional[Any] = num_mel_bins _lowercase : Any = is_training _lowercase : int = use_labels _lowercase : Dict = hidden_size _lowercase : int = num_hidden_layers _lowercase : Any = num_attention_heads _lowercase : str = intermediate_size _lowercase : Tuple = hidden_act _lowercase : Optional[int] = hidden_dropout_prob _lowercase : List[Any] = attention_probs_dropout_prob _lowercase : Dict = type_sequence_label_size _lowercase : List[str] = initializer_range _lowercase : List[Any] = scope _lowercase : str = frequency_stride _lowercase : Optional[Any] = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) _lowercase : Optional[Any] = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 _lowercase : int = (self.max_length - self.patch_size) // self.time_stride + 1 _lowercase : List[Any] = frequency_out_dimension * time_out_dimension _lowercase : Tuple = num_patches + 2 def __a ( self ): _lowercase : List[str] = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) _lowercase : Dict = None if self.use_labels: _lowercase : int = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowercase : Union[str, Any] = self.get_config() return config, input_values, labels def __a ( self ): return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): _lowercase : Dict = ASTModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() _lowercase : List[str] = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __a ( self ): _lowercase : Optional[int] = self.prepare_config_and_inputs() ( ( _lowercase ) , ( _lowercase ) , ( _lowercase ) , ) : Union[str, Any] = config_and_inputs _lowercase : Optional[int] = {'input_values': input_values} return config, inputs_dict @require_torch class lowerCAmelCase_ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCamelCase : List[Any] = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) _UpperCamelCase : Tuple = ( {"audio-classification": ASTForAudioClassification, "feature-extraction": ASTModel} if is_torch_available() else {} ) _UpperCamelCase : int = False _UpperCamelCase : Optional[Any] = False _UpperCamelCase : Any = False _UpperCamelCase : str = False def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def __a ( self ): _lowercase : Optional[Any] = ASTModelTester(self ) _lowercase : List[Any] = ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=3_7 ) def __a ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='AST does not use inputs_embeds' ) def __a ( self ): pass def __a ( self ): _lowercase , _lowercase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowercase : List[Any] = model_class(_lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _lowercase : Tuple = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCAmelCase , nn.Linear ) ) def __a ( self ): _lowercase , _lowercase : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowercase : List[Any] = model_class(_lowerCAmelCase ) _lowercase : Tuple = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowercase : List[str] = [*signature.parameters.keys()] _lowercase : Tuple = ['input_values'] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def __a ( self ): _lowercase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) @slow def __a ( self ): for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowercase : Any = ASTModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def __magic_name__ ( ) -> Any: _lowercase : List[Any] = hf_hub_download( repo_id='nielsr/audio-spectogram-transformer-checkpoint' , filename='sample_audio.flac' , repo_type='dataset' ) _lowercase , _lowercase : Tuple = torchaudio.load(SCREAMING_SNAKE_CASE ) return audio, sampling_rate @require_torch @require_torchaudio class lowerCAmelCase_ ( unittest.TestCase ): @cached_property def __a ( self ): return ( ASTFeatureExtractor.from_pretrained('MIT/ast-finetuned-audioset-10-10-0.4593' ) if is_torchaudio_available() else None ) @slow def __a ( self ): _lowercase : List[Any] = self.default_feature_extractor _lowercase : List[Any] = ASTForAudioClassification.from_pretrained('MIT/ast-finetuned-audioset-10-10-0.4593' ).to(_lowerCAmelCase ) _lowercase : List[Any] = self.default_feature_extractor _lowercase , _lowercase : Any = prepare_audio() _lowercase : int = audio.squeeze().numpy() _lowercase : List[str] = feature_extractor(_lowerCAmelCase , sampling_rate=_lowerCAmelCase , return_tensors='pt' ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): _lowercase : List[str] = model(**_lowerCAmelCase ) # verify the logits _lowercase : Any = torch.Size((1, 5_2_7) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) _lowercase : Dict = torch.tensor([-0.87_60, -7.00_42, -8.66_02] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCAmelCase , atol=1E-4 ) )
66
from PIL import Image def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Image: def brightness(SCREAMING_SNAKE_CASE ) -> float: return 128 + level + (c - 128) if not -255.0 <= level <= 255.0: raise ValueError('level must be between -255.0 (black) and 255.0 (white)' ) return img.point(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": # Load image with Image.open("image_data/lena.jpg") as img: # Change brightness to 100 UpperCamelCase = change_brightness(img, 100) brigt_img.save("image_data/lena_brightness.png", format="png")
66
1
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = {"configuration_van": ["VAN_PRETRAINED_CONFIG_ARCHIVE_MAP", "VanConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "VAN_PRETRAINED_MODEL_ARCHIVE_LIST", "VanForImageClassification", "VanModel", "VanPreTrainedModel", ] if TYPE_CHECKING: from .configuration_van import VAN_PRETRAINED_CONFIG_ARCHIVE_MAP, VanConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_van import ( VAN_PRETRAINED_MODEL_ARCHIVE_LIST, VanForImageClassification, VanModel, VanPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure)
66
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): _lowercase : List[Any] = torch.nn.Linear(1_0 , 1_0 ) _lowercase : Any = torch.optim.SGD(model.parameters() , 0.1 ) _lowercase : str = Accelerator() _lowercase : Any = accelerator.prepare(_lowerCAmelCase ) try: pickle.loads(pickle.dumps(_lowerCAmelCase ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
66
1
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_download, hf_hub_url from PIL import Image from transformers import DetaConfig, DetaForObjectDetection, DetaImageProcessor, SwinConfig from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase = logging.get_logger(__name__) def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> List[Any]: _lowercase : Optional[Any] = SwinConfig( embed_dim=192 , depths=(2, 2, 18, 2) , num_heads=(6, 12, 24, 48) , window_size=12 , out_features=['stage2', 'stage3', 'stage4'] , ) _lowercase : Any = DetaConfig( backbone_config=SCREAMING_SNAKE_CASE , num_queries=900 , encoder_ffn_dim=2_048 , decoder_ffn_dim=2_048 , num_feature_levels=5 , assign_first_stage=SCREAMING_SNAKE_CASE , with_box_refine=SCREAMING_SNAKE_CASE , two_stage=SCREAMING_SNAKE_CASE , ) # set labels _lowercase : int = 'huggingface/label-files' if "o365" in model_name: _lowercase : Any = 366 _lowercase : Dict = 'object365-id2label.json' else: _lowercase : str = 91 _lowercase : Union[str, Any] = 'coco-detection-id2label.json' _lowercase : Optional[int] = num_labels _lowercase : Dict = json.load(open(cached_download(hf_hub_url(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , repo_type='dataset' ) ) , 'r' ) ) _lowercase : List[str] = {int(SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} _lowercase : Any = idalabel _lowercase : Union[str, Any] = {v: k for k, v in idalabel.items()} return config def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> Union[str, Any]: _lowercase : Union[str, Any] = [] # stem # fmt: off rename_keys.append(('backbone.0.body.patch_embed.proj.weight', 'model.backbone.model.embeddings.patch_embeddings.projection.weight') ) rename_keys.append(('backbone.0.body.patch_embed.proj.bias', 'model.backbone.model.embeddings.patch_embeddings.projection.bias') ) rename_keys.append(('backbone.0.body.patch_embed.norm.weight', 'model.backbone.model.embeddings.norm.weight') ) rename_keys.append(('backbone.0.body.patch_embed.norm.bias', 'model.backbone.model.embeddings.norm.bias') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F"""backbone.0.body.layers.{i}.blocks.{j}.norm1.weight""", F"""model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_before.weight""") ) rename_keys.append((F"""backbone.0.body.layers.{i}.blocks.{j}.norm1.bias""", F"""model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_before.bias""") ) rename_keys.append((F"""backbone.0.body.layers.{i}.blocks.{j}.attn.relative_position_bias_table""", F"""model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table""") ) rename_keys.append((F"""backbone.0.body.layers.{i}.blocks.{j}.attn.relative_position_index""", F"""model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index""") ) rename_keys.append((F"""backbone.0.body.layers.{i}.blocks.{j}.attn.proj.weight""", F"""model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight""") ) rename_keys.append((F"""backbone.0.body.layers.{i}.blocks.{j}.attn.proj.bias""", F"""model.backbone.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias""") ) rename_keys.append((F"""backbone.0.body.layers.{i}.blocks.{j}.norm2.weight""", F"""model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_after.weight""") ) rename_keys.append((F"""backbone.0.body.layers.{i}.blocks.{j}.norm2.bias""", F"""model.backbone.model.encoder.layers.{i}.blocks.{j}.layernorm_after.bias""") ) rename_keys.append((F"""backbone.0.body.layers.{i}.blocks.{j}.mlp.fc1.weight""", F"""model.backbone.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight""") ) rename_keys.append((F"""backbone.0.body.layers.{i}.blocks.{j}.mlp.fc1.bias""", F"""model.backbone.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias""") ) rename_keys.append((F"""backbone.0.body.layers.{i}.blocks.{j}.mlp.fc2.weight""", F"""model.backbone.model.encoder.layers.{i}.blocks.{j}.output.dense.weight""") ) rename_keys.append((F"""backbone.0.body.layers.{i}.blocks.{j}.mlp.fc2.bias""", F"""model.backbone.model.encoder.layers.{i}.blocks.{j}.output.dense.bias""") ) if i < 3: rename_keys.append((F"""backbone.0.body.layers.{i}.downsample.reduction.weight""", F"""model.backbone.model.encoder.layers.{i}.downsample.reduction.weight""") ) rename_keys.append((F"""backbone.0.body.layers.{i}.downsample.norm.weight""", F"""model.backbone.model.encoder.layers.{i}.downsample.norm.weight""") ) rename_keys.append((F"""backbone.0.body.layers.{i}.downsample.norm.bias""", F"""model.backbone.model.encoder.layers.{i}.downsample.norm.bias""") ) rename_keys.append(('backbone.0.body.norm1.weight', 'model.backbone.model.hidden_states_norms.stage2.weight') ) rename_keys.append(('backbone.0.body.norm1.bias', 'model.backbone.model.hidden_states_norms.stage2.bias') ) rename_keys.append(('backbone.0.body.norm2.weight', 'model.backbone.model.hidden_states_norms.stage3.weight') ) rename_keys.append(('backbone.0.body.norm2.bias', 'model.backbone.model.hidden_states_norms.stage3.bias') ) rename_keys.append(('backbone.0.body.norm3.weight', 'model.backbone.model.hidden_states_norms.stage4.weight') ) rename_keys.append(('backbone.0.body.norm3.bias', 'model.backbone.model.hidden_states_norms.stage4.bias') ) # transformer encoder for i in range(config.encoder_layers ): rename_keys.append((F"""transformer.encoder.layers.{i}.self_attn.sampling_offsets.weight""", F"""model.encoder.layers.{i}.self_attn.sampling_offsets.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.self_attn.sampling_offsets.bias""", F"""model.encoder.layers.{i}.self_attn.sampling_offsets.bias""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.self_attn.attention_weights.weight""", F"""model.encoder.layers.{i}.self_attn.attention_weights.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.self_attn.attention_weights.bias""", F"""model.encoder.layers.{i}.self_attn.attention_weights.bias""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.self_attn.value_proj.weight""", F"""model.encoder.layers.{i}.self_attn.value_proj.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.self_attn.value_proj.bias""", F"""model.encoder.layers.{i}.self_attn.value_proj.bias""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.self_attn.output_proj.weight""", F"""model.encoder.layers.{i}.self_attn.output_proj.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.self_attn.output_proj.bias""", F"""model.encoder.layers.{i}.self_attn.output_proj.bias""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.norm1.weight""", F"""model.encoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.norm1.bias""", F"""model.encoder.layers.{i}.self_attn_layer_norm.bias""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.weight""", F"""model.encoder.layers.{i}.fc1.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.bias""", F"""model.encoder.layers.{i}.fc1.bias""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.weight""", F"""model.encoder.layers.{i}.fc2.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.bias""", F"""model.encoder.layers.{i}.fc2.bias""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.norm2.weight""", F"""model.encoder.layers.{i}.final_layer_norm.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.norm2.bias""", F"""model.encoder.layers.{i}.final_layer_norm.bias""") ) # transformer decoder for i in range(config.decoder_layers ): rename_keys.append((F"""transformer.decoder.layers.{i}.cross_attn.sampling_offsets.weight""", F"""model.decoder.layers.{i}.encoder_attn.sampling_offsets.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.cross_attn.sampling_offsets.bias""", F"""model.decoder.layers.{i}.encoder_attn.sampling_offsets.bias""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.cross_attn.attention_weights.weight""", F"""model.decoder.layers.{i}.encoder_attn.attention_weights.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.cross_attn.attention_weights.bias""", F"""model.decoder.layers.{i}.encoder_attn.attention_weights.bias""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.cross_attn.value_proj.weight""", F"""model.decoder.layers.{i}.encoder_attn.value_proj.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.cross_attn.value_proj.bias""", F"""model.decoder.layers.{i}.encoder_attn.value_proj.bias""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.cross_attn.output_proj.weight""", F"""model.decoder.layers.{i}.encoder_attn.output_proj.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.cross_attn.output_proj.bias""", F"""model.decoder.layers.{i}.encoder_attn.output_proj.bias""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm1.weight""", F"""model.decoder.layers.{i}.encoder_attn_layer_norm.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm1.bias""", F"""model.decoder.layers.{i}.encoder_attn_layer_norm.bias""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.self_attn.out_proj.weight""", F"""model.decoder.layers.{i}.self_attn.out_proj.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.self_attn.out_proj.bias""", F"""model.decoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm2.weight""", F"""model.decoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm2.bias""", F"""model.decoder.layers.{i}.self_attn_layer_norm.bias""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.weight""", F"""model.decoder.layers.{i}.fc1.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.bias""", F"""model.decoder.layers.{i}.fc1.bias""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.weight""", F"""model.decoder.layers.{i}.fc2.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.bias""", F"""model.decoder.layers.{i}.fc2.bias""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm3.weight""", F"""model.decoder.layers.{i}.final_layer_norm.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm3.bias""", F"""model.decoder.layers.{i}.final_layer_norm.bias""") ) # fmt: on return rename_keys def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> List[str]: _lowercase : Any = dct.pop(SCREAMING_SNAKE_CASE ) _lowercase : List[str] = val def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Union[str, Any]: _lowercase : Optional[Any] = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): _lowercase : Union[str, Any] = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) _lowercase : str = state_dict.pop(F"""backbone.0.body.layers.{i}.blocks.{j}.attn.qkv.weight""" ) _lowercase : Optional[Any] = state_dict.pop(F"""backbone.0.body.layers.{i}.blocks.{j}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict _lowercase : List[str] = in_proj_weight[:dim, :] _lowercase : List[str] = in_proj_bias[: dim] _lowercase : str = in_proj_weight[ dim : dim * 2, : ] _lowercase : Union[str, Any] = in_proj_bias[ dim : dim * 2 ] _lowercase : Dict = in_proj_weight[ -dim :, : ] _lowercase : Union[str, Any] = in_proj_bias[-dim :] # fmt: on def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Optional[Any]: # transformer decoder self-attention layers _lowercase : str = config.d_model for i in range(config.decoder_layers ): # read in weights + bias of input projection layer of self-attention _lowercase : Dict = state_dict.pop(F"""transformer.decoder.layers.{i}.self_attn.in_proj_weight""" ) _lowercase : str = state_dict.pop(F"""transformer.decoder.layers.{i}.self_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict _lowercase : List[Any] = in_proj_weight[:hidden_size, :] _lowercase : Optional[Any] = in_proj_bias[:hidden_size] _lowercase : List[str] = in_proj_weight[ hidden_size : hidden_size * 2, : ] _lowercase : str = in_proj_bias[hidden_size : hidden_size * 2] _lowercase : str = in_proj_weight[-hidden_size:, :] _lowercase : Dict = in_proj_bias[-hidden_size:] def __magic_name__ ( ) -> Optional[int]: _lowercase : str = 'http://images.cocodataset.org/val2017/000000039769.jpg' _lowercase : List[Any] = Image.open(requests.get(SCREAMING_SNAKE_CASE , stream=SCREAMING_SNAKE_CASE ).raw ) return im @torch.no_grad() def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Optional[Any]: _lowercase : Dict = get_deta_config(SCREAMING_SNAKE_CASE ) # load original state dict if model_name == "deta-swin-large": _lowercase : int = hf_hub_download(repo_id='nielsr/deta-checkpoints' , filename='adet_swin_ft.pth' ) elif model_name == "deta-swin-large-o365": _lowercase : Dict = hf_hub_download(repo_id='jozhang97/deta-swin-l-o365' , filename='deta_swin_pt_o365.pth' ) else: raise ValueError(F"""Model name {model_name} not supported""" ) _lowercase : str = torch.load(SCREAMING_SNAKE_CASE , map_location='cpu' )['model'] # original state dict for name, param in state_dict.items(): print(SCREAMING_SNAKE_CASE , param.shape ) # rename keys _lowercase : Tuple = create_rename_keys(SCREAMING_SNAKE_CASE ) for src, dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) read_in_swin_q_k_v(SCREAMING_SNAKE_CASE , config.backbone_config ) read_in_decoder_q_k_v(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # fix some prefixes for key in state_dict.copy().keys(): if "transformer.decoder.class_embed" in key or "transformer.decoder.bbox_embed" in key: _lowercase : int = state_dict.pop(SCREAMING_SNAKE_CASE ) _lowercase : List[str] = val if "input_proj" in key: _lowercase : Any = state_dict.pop(SCREAMING_SNAKE_CASE ) _lowercase : int = val if "level_embed" in key or "pos_trans" in key or "pix_trans" in key or "enc_output" in key: _lowercase : Union[str, Any] = state_dict.pop(SCREAMING_SNAKE_CASE ) _lowercase : Dict = val # finally, create HuggingFace model and load state dict _lowercase : Optional[Any] = DetaForObjectDetection(SCREAMING_SNAKE_CASE ) model.load_state_dict(SCREAMING_SNAKE_CASE ) model.eval() _lowercase : Tuple = 'cuda' if torch.cuda.is_available() else 'cpu' model.to(SCREAMING_SNAKE_CASE ) # load image processor _lowercase : Optional[int] = DetaImageProcessor(format='coco_detection' ) # verify our conversion on image _lowercase : int = prepare_img() _lowercase : Tuple = processor(images=SCREAMING_SNAKE_CASE , return_tensors='pt' ) _lowercase : Union[str, Any] = encoding['pixel_values'] _lowercase : Any = model(pixel_values.to(SCREAMING_SNAKE_CASE ) ) # verify logits print('Logits:' , outputs.logits[0, :3, :3] ) print('Boxes:' , outputs.pred_boxes[0, :3, :3] ) if model_name == "deta-swin-large": _lowercase : Tuple = torch.tensor( [[-7.6308, -2.8485, -5.3737], [-7.2037, -4.5505, -4.8027], [-7.2943, -4.2611, -4.6617]] ) _lowercase : Dict = torch.tensor([[0.4987, 0.4969, 0.9999], [0.2549, 0.5498, 0.4805], [0.5498, 0.2757, 0.0569]] ) elif model_name == "deta-swin-large-o365": _lowercase : Union[str, Any] = torch.tensor( [[-8.0122, -3.5720, -4.9717], [-8.1547, -3.6886, -4.6389], [-7.6610, -3.6194, -5.0134]] ) _lowercase : Dict = torch.tensor([[0.2523, 0.5549, 0.4881], [0.7715, 0.4149, 0.4601], [0.5503, 0.2753, 0.0575]] ) assert torch.allclose(outputs.logits[0, :3, :3] , expected_logits.to(SCREAMING_SNAKE_CASE ) , atol=1E-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , expected_boxes.to(SCREAMING_SNAKE_CASE ) , atol=1E-4 ) print('Everything ok!' ) if pytorch_dump_folder_path: # Save model and processor logger.info(F"""Saving PyTorch model and processor to {pytorch_dump_folder_path}...""" ) Path(SCREAMING_SNAKE_CASE ).mkdir(exist_ok=SCREAMING_SNAKE_CASE ) model.save_pretrained(SCREAMING_SNAKE_CASE ) processor.save_pretrained(SCREAMING_SNAKE_CASE ) # Push to hub if push_to_hub: print('Pushing model and processor to hub...' ) model.push_to_hub(F"""jozhang97/{model_name}""" ) processor.push_to_hub(F"""jozhang97/{model_name}""" ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() parser.add_argument( "--model_name", type=str, default="deta-swin-large", choices=["deta-swin-large", "deta-swin-large-o365"], help="Name of the model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the folder to output PyTorch model.", ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) UpperCamelCase = parser.parse_args() convert_deta_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
66
import requests from bsa import BeautifulSoup def __magic_name__ ( SCREAMING_SNAKE_CASE = "AAPL" ) -> str: _lowercase : str = F"""https://in.finance.yahoo.com/quote/{symbol}?s={symbol}""" _lowercase : int = BeautifulSoup(requests.get(SCREAMING_SNAKE_CASE ).text , 'html.parser' ) _lowercase : List[str] = 'My(6px) Pos(r) smartphone_Mt(6px)' return soup.find('div' , class_=class_ ).find('span' ).text if __name__ == "__main__": for symbol in "AAPL AMZN IBM GOOG MSFT ORCL".split(): print(f'''Current {symbol:<4} stock price is {stock_price(symbol):>8}''')
66
1
import os from datetime import datetime as dt from github import Github UpperCamelCase = [ "good first issue", "feature request", "wip", ] def __magic_name__ ( ) -> List[str]: _lowercase : Union[str, Any] = Github(os.environ['GITHUB_TOKEN'] ) _lowercase : str = g.get_repo('huggingface/accelerate' ) _lowercase : Any = repo.get_issues(state='open' ) for issue in open_issues: _lowercase : Dict = sorted([comment for comment in issue.get_comments()] , key=lambda SCREAMING_SNAKE_CASE : i.created_at , reverse=SCREAMING_SNAKE_CASE ) _lowercase : Any = comments[0] if len(SCREAMING_SNAKE_CASE ) > 0 else None _lowercase : Optional[Any] = dt.utcnow() _lowercase : str = (current_time - issue.updated_at).days _lowercase : Dict = (current_time - issue.created_at).days if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and days_since_updated > 7 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Close issue since it has been 7 days of inactivity since bot mention. issue.edit(state='closed' ) elif ( days_since_updated > 23 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Add stale comment issue.create_comment( 'This issue has been automatically marked as stale because it has not had ' 'recent activity. If you think this still needs to be addressed ' 'please comment on this thread.\n\nPlease note that issues that do not follow the ' '[contributing guidelines](https://github.com/huggingface/accelerate/blob/main/CONTRIBUTING.md) ' 'are likely to be ignored.' ) if __name__ == "__main__": main()
66
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = { "configuration_poolformer": [ "POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "PoolFormerConfig", "PoolFormerOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["PoolFormerFeatureExtractor"] UpperCamelCase = ["PoolFormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "PoolFormerForImageClassification", "PoolFormerModel", "PoolFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure)
66
1
from typing import Any def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> list[Any]: if not input_list: return [] _lowercase : Dict = [input_list.count(SCREAMING_SNAKE_CASE ) for value in input_list] _lowercase : List[Any] = max(SCREAMING_SNAKE_CASE ) # Gets the maximum count in the input list. # Gets values of modes return sorted({input_list[i] for i, value in enumerate(SCREAMING_SNAKE_CASE ) if value == y} ) if __name__ == "__main__": import doctest doctest.testmod()
66
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : Dict = LayoutLMTokenizer _UpperCamelCase : Union[str, Any] = LayoutLMTokenizerFast _UpperCamelCase : int = True _UpperCamelCase : Tuple = True def __a ( self ): super().setUp() _lowercase : Union[str, Any] = [ '[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] _lowercase : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def __a ( self , **_lowerCAmelCase ): return LayoutLMTokenizer.from_pretrained(self.tmpdirname , **_lowerCAmelCase ) def __a ( self , _lowerCAmelCase ): _lowercase : str = 'UNwant\u00E9d,running' _lowercase : List[Any] = 'unwanted, running' return input_text, output_text def __a ( self ): _lowercase : Dict = self.tokenizer_class(self.vocab_file ) _lowercase : Dict = tokenizer.tokenize('UNwant\u00E9d,running' ) self.assertListEqual(_lowerCAmelCase , ['un', '##want', '##ed', ',', 'runn', '##ing'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [7, 4, 5, 1_0, 8, 9] ) def __a ( self ): pass
66
1
import copy from typing import Dict, List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING UpperCamelCase = { "facebook/mask2former-swin-small-coco-instance": ( "https://huggingface.co/facebook/mask2former-swin-small-coco-instance/blob/main/config.json" ) # See all Mask2Former models at https://huggingface.co/models?filter=mask2former } UpperCamelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : Dict = "mask2former" _UpperCamelCase : int = ["swin"] _UpperCamelCase : str = {"hidden_size": "hidden_dim"} def __init__( self , _lowerCAmelCase = None , _lowerCAmelCase = 2_5_6 , _lowerCAmelCase = 2_5_6 , _lowerCAmelCase = 2_5_6 , _lowerCAmelCase = 1_0_2_4 , _lowerCAmelCase = "relu" , _lowerCAmelCase = 6 , _lowerCAmelCase = 1_0 , _lowerCAmelCase = 8 , _lowerCAmelCase = 0.0 , _lowerCAmelCase = 2_0_4_8 , _lowerCAmelCase = False , _lowerCAmelCase = False , _lowerCAmelCase = 4 , _lowerCAmelCase = 2_5_5 , _lowerCAmelCase = 1_0_0 , _lowerCAmelCase = 0.1 , _lowerCAmelCase = 2.0 , _lowerCAmelCase = 5.0 , _lowerCAmelCase = 5.0 , _lowerCAmelCase = 1_2_5_4_4 , _lowerCAmelCase = 3.0 , _lowerCAmelCase = 0.75 , _lowerCAmelCase = 0.02 , _lowerCAmelCase = 1.0 , _lowerCAmelCase = True , _lowerCAmelCase = [4, 8, 1_6, 3_2] , _lowerCAmelCase = None , **_lowerCAmelCase , ): if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `Swin` backbone.' ) _lowercase : int = CONFIG_MAPPING['swin']( image_size=2_2_4 , in_channels=3 , patch_size=4 , embed_dim=9_6 , depths=[2, 2, 1_8, 2] , num_heads=[3, 6, 1_2, 2_4] , window_size=7 , drop_path_rate=0.3 , use_absolute_embeddings=_lowerCAmelCase , out_features=['stage1', 'stage2', 'stage3', 'stage4'] , ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ): _lowercase : Any = backbone_config.pop('model_type' ) _lowercase : Any = CONFIG_MAPPING[backbone_model_type] _lowercase : Any = config_class.from_dict(_lowerCAmelCase ) # verify that the backbone is supported if backbone_config.model_type not in self.backbones_supported: logger.warning_once( F"""Backbone {backbone_config.model_type} is not a supported model and may not be compatible with Mask2Former. """ F"""Supported model types: {','.join(self.backbones_supported )}""" ) _lowercase : Optional[Any] = backbone_config _lowercase : str = feature_size _lowercase : Any = mask_feature_size _lowercase : List[Any] = hidden_dim _lowercase : int = encoder_feedforward_dim _lowercase : str = activation_function _lowercase : int = encoder_layers _lowercase : str = decoder_layers _lowercase : str = num_attention_heads _lowercase : List[str] = dropout _lowercase : Tuple = dim_feedforward _lowercase : int = pre_norm _lowercase : int = enforce_input_projection _lowercase : Optional[Any] = common_stride _lowercase : Union[str, Any] = ignore_value _lowercase : Union[str, Any] = num_queries _lowercase : Optional[Any] = no_object_weight _lowercase : List[Any] = class_weight _lowercase : Tuple = mask_weight _lowercase : Optional[int] = dice_weight _lowercase : Any = train_num_points _lowercase : List[Any] = oversample_ratio _lowercase : Dict = importance_sample_ratio _lowercase : Union[str, Any] = init_std _lowercase : int = init_xavier_std _lowercase : List[Any] = use_auxiliary_loss _lowercase : Optional[Any] = feature_strides _lowercase : Dict = output_auxiliary_logits _lowercase : str = decoder_layers super().__init__(**_lowerCAmelCase ) @classmethod def __a ( cls , _lowerCAmelCase , **_lowerCAmelCase ): return cls( backbone_config=_lowerCAmelCase , **_lowerCAmelCase , ) def __a ( self ): _lowercase : Union[str, Any] = copy.deepcopy(self.__dict__ ) _lowercase : str = self.backbone_config.to_dict() _lowercase : Tuple = self.__class__.model_type return output
66
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : str = ShapEPipeline _UpperCamelCase : Any = ["prompt"] _UpperCamelCase : int = ["prompt"] _UpperCamelCase : Union[str, Any] = [ "num_images_per_prompt", "num_inference_steps", "generator", "latents", "guidance_scale", "frame_size", "output_type", "return_dict", ] _UpperCamelCase : Optional[Any] = False @property def __a ( self ): return 3_2 @property def __a ( self ): return 3_2 @property def __a ( self ): return self.time_input_dim * 4 @property def __a ( self ): return 8 @property def __a ( self ): _lowercase : Any = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) return tokenizer @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) return CLIPTextModelWithProjection(_lowerCAmelCase ) @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = { 'num_attention_heads': 2, 'attention_head_dim': 1_6, 'embedding_dim': self.time_input_dim, 'num_embeddings': 3_2, 'embedding_proj_dim': self.text_embedder_hidden_size, 'time_embed_dim': self.time_embed_dim, 'num_layers': 1, 'clip_embed_dim': self.time_input_dim * 2, 'additional_embeddings': 0, 'time_embed_act_fn': 'gelu', 'norm_in_type': 'layer', 'encoder_hid_proj_type': None, 'added_emb_type': None, } _lowercase : Optional[Any] = PriorTransformer(**_lowerCAmelCase ) return model @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = { 'param_shapes': ( (self.renderer_dim, 9_3), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), 'd_latent': self.time_input_dim, 'd_hidden': self.renderer_dim, 'n_output': 1_2, 'background': ( 0.1, 0.1, 0.1, ), } _lowercase : List[Any] = ShapERenderer(**_lowerCAmelCase ) return model def __a ( self ): _lowercase : Optional[Any] = self.dummy_prior _lowercase : Dict = self.dummy_text_encoder _lowercase : List[str] = self.dummy_tokenizer _lowercase : Union[str, Any] = self.dummy_renderer _lowercase : List[str] = HeunDiscreteScheduler( beta_schedule='exp' , num_train_timesteps=1_0_2_4 , prediction_type='sample' , use_karras_sigmas=_lowerCAmelCase , clip_sample=_lowerCAmelCase , clip_sample_range=1.0 , ) _lowercase : List[str] = { 'prior': prior, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'renderer': renderer, 'scheduler': scheduler, } return components def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : Optional[Any] = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Optional[int] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : List[Any] = { 'prompt': 'horse', 'generator': generator, 'num_inference_steps': 1, 'frame_size': 3_2, 'output_type': 'np', } return inputs def __a ( self ): _lowercase : Optional[int] = 'cpu' _lowercase : List[Any] = self.get_dummy_components() _lowercase : Tuple = self.pipeline_class(**_lowerCAmelCase ) _lowercase : Union[str, Any] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Union[str, Any] = pipe(**self.get_dummy_inputs(_lowerCAmelCase ) ) _lowercase : str = output.images[0] _lowercase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (2_0, 3_2, 3_2, 3) _lowercase : str = np.array( [ 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __a ( self ): # NOTE: Larger batch sizes cause this test to timeout, only test on smaller batches self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def __a ( self ): _lowercase : List[Any] = torch_device == 'cpu' _lowercase : Any = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=_lowerCAmelCase , relax_max_difference=_lowerCAmelCase , ) def __a ( self ): _lowercase : Union[str, Any] = self.get_dummy_components() _lowercase : Optional[int] = self.pipeline_class(**_lowerCAmelCase ) _lowercase : Any = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : str = 1 _lowercase : Optional[int] = 2 _lowercase : List[str] = self.get_dummy_inputs(_lowerCAmelCase ) for key in inputs.keys(): if key in self.batch_params: _lowercase : int = batch_size * [inputs[key]] _lowercase : Optional[int] = pipe(**_lowerCAmelCase , num_images_per_prompt=_lowerCAmelCase )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ): _lowercase : List[str] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/shap_e/test_shap_e_np_out.npy' ) _lowercase : Any = ShapEPipeline.from_pretrained('openai/shap-e' ) _lowercase : List[str] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Tuple = torch.Generator(device=_lowerCAmelCase ).manual_seed(0 ) _lowercase : int = pipe( 'a shark' , generator=_lowerCAmelCase , guidance_scale=15.0 , num_inference_steps=6_4 , frame_size=6_4 , output_type='np' , ).images[0] assert images.shape == (2_0, 6_4, 6_4, 3) assert_mean_pixel_difference(_lowerCAmelCase , _lowerCAmelCase )
66
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) UpperCamelCase = { "configuration_mobilevit": ["MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "MobileViTConfig", "MobileViTOnnxConfig"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["MobileViTFeatureExtractor"] UpperCamelCase = ["MobileViTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "MobileViTForImageClassification", "MobileViTForSemanticSegmentation", "MobileViTModel", "MobileViTPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFMobileViTForImageClassification", "TFMobileViTForSemanticSegmentation", "TFMobileViTModel", "TFMobileViTPreTrainedModel", ] if TYPE_CHECKING: from .configuration_mobilevit import MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileViTConfig, MobileViTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_mobilevit import MobileViTFeatureExtractor from .image_processing_mobilevit import MobileViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilevit import ( MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel, MobileViTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilevit import ( TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileViTForImageClassification, TFMobileViTForSemanticSegmentation, TFMobileViTModel, TFMobileViTPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
import sys UpperCamelCase = ( "73167176531330624919225119674426574742355349194934" "96983520312774506326239578318016984801869478851843" "85861560789112949495459501737958331952853208805511" "12540698747158523863050715693290963295227443043557" "66896648950445244523161731856403098711121722383113" "62229893423380308135336276614282806444486645238749" "30358907296290491560440772390713810515859307960866" "70172427121883998797908792274921901699720888093776" "65727333001053367881220235421809751254540594752243" "52584907711670556013604839586446706324415722155397" "53697817977846174064955149290862569321978468622482" "83972241375657056057490261407972968652414535100474" "82166370484403199890008895243450658541227588666881" "16427171479924442928230863465674813919123162824586" "17866458359124566529476545682848912883142607690042" "24219022671055626321111109370544217506941658960408" "07198403850962455444362981230987879927244284909188" "84580156166097919133875499200524063689912560717606" "05886116467109405077541002256983155200055935729725" "71636269561882670428252483600823257530420752963450" ) def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : List[Any] = 1 for digit in s: product *= int(SCREAMING_SNAKE_CASE ) return product def __magic_name__ ( SCREAMING_SNAKE_CASE = N ) -> int: _lowercase : Dict = -sys.maxsize - 1 _lowercase : Tuple = n[:13] _lowercase : List[Any] = 13 while cur_index < len(SCREAMING_SNAKE_CASE ) - 13: if int(n[cur_index] ) >= int(substr[0] ): _lowercase : List[str] = substr[1:] + n[cur_index] cur_index += 1 else: _lowercase : str = max(SCREAMING_SNAKE_CASE , str_eval(SCREAMING_SNAKE_CASE ) ) _lowercase : Dict = n[cur_index : cur_index + 13] cur_index += 13 return largest_product if __name__ == "__main__": print(f'''{solution() = }''')
66
1
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> list: _lowercase : Dict = len(SCREAMING_SNAKE_CASE ) _lowercase : Any = [[0] * n for i in range(SCREAMING_SNAKE_CASE )] for i in range(SCREAMING_SNAKE_CASE ): _lowercase : Optional[Any] = y_points[i] for i in range(2 , SCREAMING_SNAKE_CASE ): for j in range(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): _lowercase : Union[str, Any] = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
66
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available UpperCamelCase = { "configuration_gpt_neo": ["GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTNeoConfig", "GPTNeoOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTNeoForCausalLM", "GPTNeoForQuestionAnswering", "GPTNeoForSequenceClassification", "GPTNeoForTokenClassification", "GPTNeoModel", "GPTNeoPreTrainedModel", "load_tf_weights_in_gpt_neo", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "FlaxGPTNeoForCausalLM", "FlaxGPTNeoModel", "FlaxGPTNeoPreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_neo import GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoConfig, GPTNeoOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neo import ( GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoForCausalLM, GPTNeoForQuestionAnswering, GPTNeoForSequenceClassification, GPTNeoForTokenClassification, GPTNeoModel, GPTNeoPreTrainedModel, load_tf_weights_in_gpt_neo, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_gpt_neo import FlaxGPTNeoForCausalLM, FlaxGPTNeoModel, FlaxGPTNeoPreTrainedModel else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
1
import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": UpperCamelCase = pd.read_csv("sample_data.csv", header=None) UpperCamelCase = df.shape[:1][0] # If you're using some other dataset input the target column UpperCamelCase = df.iloc[:, 1:2] UpperCamelCase = actual_data.values.reshape(len_data, 1) UpperCamelCase = MinMaxScaler().fit_transform(actual_data) UpperCamelCase = 10 UpperCamelCase = 5 UpperCamelCase = 20 UpperCamelCase = len_data - periods * look_back UpperCamelCase = actual_data[:division] UpperCamelCase = actual_data[division - look_back :] UpperCamelCase , UpperCamelCase = [], [] UpperCamelCase , UpperCamelCase = [], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) UpperCamelCase = np.array(train_x) UpperCamelCase = np.array(test_x) UpperCamelCase = np.array([list(i.ravel()) for i in train_y]) UpperCamelCase = np.array([list(i.ravel()) for i in test_y]) UpperCamelCase = Sequential() model.add(LSTM(128, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(64, input_shape=(128, 1))) model.add(Dense(forward_days)) model.compile(loss="mean_squared_error", optimizer="adam") UpperCamelCase = model.fit( x_train, y_train, epochs=150, verbose=1, shuffle=True, batch_size=4 ) UpperCamelCase = model.predict(x_test)
66
from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : str = ["image_processor", "tokenizer"] _UpperCamelCase : Union[str, Any] = "AutoImageProcessor" _UpperCamelCase : Union[str, Any] = "AutoTokenizer" def __init__( self , _lowerCAmelCase , _lowerCAmelCase ): super().__init__(_lowerCAmelCase , _lowerCAmelCase ) _lowercase : Union[str, Any] = self.image_processor def __call__( self , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=None , **_lowerCAmelCase ): if text is None and images is None: raise ValueError('You have to specify either text or images. Both cannot be none.' ) if text is not None: _lowercase : Dict = self.tokenizer(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if images is not None: _lowercase : Union[str, Any] = self.image_processor(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if text is not None and images is not None: _lowercase : Optional[Any] = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_lowerCAmelCase ) , tensor_type=_lowerCAmelCase ) def __a ( self , *_lowerCAmelCase , **_lowerCAmelCase ): return self.tokenizer.batch_decode(*_lowerCAmelCase , **_lowerCAmelCase ) def __a ( self , *_lowerCAmelCase , **_lowerCAmelCase ): return self.tokenizer.decode(*_lowerCAmelCase , **_lowerCAmelCase ) @property def __a ( self ): return ["input_ids", "attention_mask", "pixel_values"]
66
1
import unittest from transformers import SPIECE_UNDERLINE, XLNetTokenizer, XLNetTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin UpperCamelCase = get_tests_dir("fixtures/test_sentencepiece.model") @require_sentencepiece @require_tokenizers class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : Dict = XLNetTokenizer _UpperCamelCase : str = XLNetTokenizerFast _UpperCamelCase : Optional[Any] = True _UpperCamelCase : int = True def __a ( self ): super().setUp() # We have a SentencePiece fixture for testing _lowercase : int = XLNetTokenizer(_lowerCAmelCase , keep_accents=_lowerCAmelCase ) tokenizer.sanitize_special_tokens() tokenizer.save_pretrained(self.tmpdirname ) def __a ( self ): _lowercase : Optional[Any] = '<s>' _lowercase : Any = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_lowerCAmelCase ) , _lowerCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_lowerCAmelCase ) , _lowerCAmelCase ) def __a ( self ): _lowercase : Tuple = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<unk>' ) self.assertEqual(vocab_keys[1] , '<s>' ) self.assertEqual(vocab_keys[-1] , '<eod>' ) self.assertEqual(len(_lowerCAmelCase ) , 1_0_0_6 ) def __a ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1_0_0_0 ) def __a ( self ): _lowercase : str = XLNetTokenizer(_lowerCAmelCase , keep_accents=_lowerCAmelCase ) _lowercase : str = tokenizer.tokenize('This is a test' ) self.assertListEqual(_lowerCAmelCase , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2] ) _lowercase : Optional[int] = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) _lowercase : Any = tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 0, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 0, 4] ) _lowercase : List[str] = tokenizer.convert_ids_to_tokens(_lowerCAmelCase ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) def __a ( self ): _lowercase : List[str] = XLNetTokenizer(_lowerCAmelCase , do_lower_case=_lowerCAmelCase ) _lowercase : Union[str, Any] = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + '', 'i', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 'se', '.', ] , ) self.assertListEqual(tokenizer.tokenize('H\u00E9llo' ) , ['▁he', 'll', 'o'] ) def __a ( self ): _lowercase : Union[str, Any] = XLNetTokenizer(_lowerCAmelCase , do_lower_case=_lowerCAmelCase ) _lowercase : Optional[int] = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 'se', '.', ] , ) @slow def __a ( self ): _lowercase : int = XLNetTokenizer.from_pretrained('xlnet-base-cased' ) _lowercase : int = tokenizer.encode('sequence builders' , add_special_tokens=_lowerCAmelCase ) _lowercase : List[str] = tokenizer.encode('multi-sequence build' , add_special_tokens=_lowerCAmelCase ) _lowercase : List[str] = tokenizer.build_inputs_with_special_tokens(_lowerCAmelCase ) _lowercase : Optional[Any] = tokenizer.build_inputs_with_special_tokens(_lowerCAmelCase , _lowerCAmelCase ) assert encoded_sentence == text + [4, 3] assert encoded_pair == text + [4] + text_a + [4, 3] @slow def __a ( self ): # fmt: off _lowercase : Any = {'input_ids': [[1_7, 2_1_4_4_2, 2_7_0, 1_7, 1_0, 1_4_6_4_5, 3_1_8, 3_4, 1_7, 4_5_4_6, 3_1_4_5, 7_8_7, 1_3, 7_7_5_2, 2_2_0_1_8, 2_3, 2_1, 1_7, 4_5_4_6, 3_1_4_5, 7_8_7, 1_3, 3_3_5_2, 1_4_4_3_1, 1_3, 5_5_0_0, 1_1, 1_1_7_6, 5_8_0, 1_3, 1_6_8_1_9, 4_7_9_7, 2_3, 1_7, 1_0, 1_7_1_3_5, 6_5_8, 1_9, 4_5_7, 7_9_3_2, 1_3, 1_8_4, 1_9, 3_1_5_4, 1_7_1_3_5, 6_4_6_8, 1_9, 1_4_0_4, 1_2_2_6_9, 1_9, 4_2_2_9, 5_3_5_6, 1_6_2_6_4, 4_6, 1_9, 1_7, 2_0_5_4_5, 1_0_3_9_5, 9, 9, 9, 1_1, 2_8, 6_4_2_1, 9_5_3_1, 2_0_7_2_9, 1_7, 1_0, 3_5_3, 1_7_0_2_2, 1_1, 2_1, 6_4_2_1, 9_5_3_1, 1_6_9_4_9, 1_7, 1_0, 1_1_5_0_9, 7_5_3, 1_1, 3_3, 9_5, 2_4_2_1, 7_3_8_5, 9_5_6, 1_4_4_3_1, 2_6_2_6, 2_5, 8_4_2, 7_3_8_5, 4_8_3_6, 2_1, 1_4_2_9, 2_2_7_2, 9_8_5_5, 3_1_2_0, 1_6_1, 2_4_7_3_8, 1_9, 1_3_2_0_3, 6_5_8, 2_1_8, 7_8_7, 2_1, 4_3_0, 1_8_4_8_2, 8_4_7, 2_6_3_7, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 3_2_2, 2_2_1_7_8, 2_7, 1_0_6_4, 2_2, 9_5_6, 1_3, 1_1_1_0_1, 1_4_2_9, 5_8_5_4, 2_4_3_1_3, 1_8_9_5_3, 4_0, 4_2_2, 2_4_3_6_6, 6_8, 1_7_5_8, 3_7, 1_0_4_8_3, 1_4_2_5_7, 3_1, 2_0_7, 2_6_3, 2_1, 2_0_3, 3_7_7_3, 2_5, 7_1, 9_7_3_5, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 3_2, 2_0_4_9, 3_4_4_2, 1_7, 1_3_8_9_4, 3_3_8_0, 2_3, 9_5, 1_8, 1_7_6_3_4, 2_2_8_8, 9, 4, 3]], 'token_type_ids': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_lowerCAmelCase , model_name='xlnet-base-cased' , revision='c841166438c31ec7ca9a106dee7bb312b73ae511' , )
66
from __future__ import annotations import math def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> list[int]: if num <= 0: _lowercase : List[str] = F"""{num}: Invalid input, please enter a positive integer.""" raise ValueError(SCREAMING_SNAKE_CASE ) _lowercase : Union[str, Any] = [True] * (num + 1) _lowercase : Union[str, Any] = [] _lowercase : Dict = 2 _lowercase : Union[str, Any] = int(math.sqrt(SCREAMING_SNAKE_CASE ) ) while start <= end: # If start is a prime if sieve[start] is True: prime.append(SCREAMING_SNAKE_CASE ) # Set multiples of start be False for i in range(start * start , num + 1 , SCREAMING_SNAKE_CASE ): if sieve[i] is True: _lowercase : str = False start += 1 for j in range(end + 1 , num + 1 ): if sieve[j] is True: prime.append(SCREAMING_SNAKE_CASE ) return prime if __name__ == "__main__": print(prime_sieve(int(input("Enter a positive integer: ").strip())))
66
1
from typing import Dict from .base import GenericTensor, Pipeline class lowerCAmelCase_ ( __snake_case ): def __a ( self , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=None , **_lowerCAmelCase ): if tokenize_kwargs is None: _lowercase : Union[str, Any] = {} if truncation is not None: if "truncation" in tokenize_kwargs: raise ValueError( 'truncation parameter defined twice (given as keyword argument as well as in tokenize_kwargs)' ) _lowercase : List[str] = truncation _lowercase : str = tokenize_kwargs _lowercase : Union[str, Any] = {} if return_tensors is not None: _lowercase : List[str] = return_tensors return preprocess_params, {}, postprocess_params def __a ( self , _lowerCAmelCase , **_lowerCAmelCase ): _lowercase : List[str] = self.framework _lowercase : Dict = self.tokenizer(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) return model_inputs def __a ( self , _lowerCAmelCase ): _lowercase : Tuple = self.model(**_lowerCAmelCase ) return model_outputs def __a ( self , _lowerCAmelCase , _lowerCAmelCase=False ): # [0] is the first available tensor, logits or last_hidden_state. if return_tensors: return model_outputs[0] if self.framework == "pt": return model_outputs[0].tolist() elif self.framework == "tf": return model_outputs[0].numpy().tolist() def __call__( self , *_lowerCAmelCase , **_lowerCAmelCase ): return super().__call__(*_lowerCAmelCase , **_lowerCAmelCase )
66
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ConvNextConfig, SegformerImageProcessor, UperNetConfig, UperNetForSemanticSegmentation def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> List[Any]: _lowercase : int = 384 if "tiny" in model_name: _lowercase : Tuple = [3, 3, 9, 3] _lowercase : List[str] = [96, 192, 384, 768] if "small" in model_name: _lowercase : List[str] = [3, 3, 27, 3] _lowercase : Union[str, Any] = [96, 192, 384, 768] if "base" in model_name: _lowercase : List[Any] = [3, 3, 27, 3] _lowercase : Dict = [128, 256, 512, 1_024] _lowercase : Optional[int] = 512 if "large" in model_name: _lowercase : List[str] = [3, 3, 27, 3] _lowercase : List[Any] = [192, 384, 768, 1_536] _lowercase : Tuple = 768 if "xlarge" in model_name: _lowercase : str = [3, 3, 27, 3] _lowercase : List[str] = [256, 512, 1_024, 2_048] _lowercase : Tuple = 1_024 # set label information _lowercase : Dict = 150 _lowercase : Union[str, Any] = 'huggingface/label-files' _lowercase : str = 'ade20k-id2label.json' _lowercase : List[Any] = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , repo_type='dataset' ) , 'r' ) ) _lowercase : Dict = {int(SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} _lowercase : Tuple = {v: k for k, v in idalabel.items()} _lowercase : List[str] = ConvNextConfig( depths=SCREAMING_SNAKE_CASE , hidden_sizes=SCREAMING_SNAKE_CASE , out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) _lowercase : Union[str, Any] = UperNetConfig( backbone_config=SCREAMING_SNAKE_CASE , auxiliary_in_channels=SCREAMING_SNAKE_CASE , num_labels=SCREAMING_SNAKE_CASE , idalabel=SCREAMING_SNAKE_CASE , labelaid=SCREAMING_SNAKE_CASE , ) return config def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : Any = [] # fmt: off # stem rename_keys.append(('backbone.downsample_layers.0.0.weight', 'backbone.embeddings.patch_embeddings.weight') ) rename_keys.append(('backbone.downsample_layers.0.0.bias', 'backbone.embeddings.patch_embeddings.bias') ) rename_keys.append(('backbone.downsample_layers.0.1.weight', 'backbone.embeddings.layernorm.weight') ) rename_keys.append(('backbone.downsample_layers.0.1.bias', 'backbone.embeddings.layernorm.bias') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F"""backbone.stages.{i}.{j}.gamma""", F"""backbone.encoder.stages.{i}.layers.{j}.layer_scale_parameter""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.depthwise_conv.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.dwconv.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.depthwise_conv.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.dwconv.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.norm.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.layernorm.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.norm.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.layernorm.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv1.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv1.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv1.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv1.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv2.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv2.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv2.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv2.bias""") ) if i > 0: rename_keys.append((F"""backbone.downsample_layers.{i}.0.weight""", F"""backbone.encoder.stages.{i}.downsampling_layer.0.weight""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.0.bias""", F"""backbone.encoder.stages.{i}.downsampling_layer.0.bias""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.1.weight""", F"""backbone.encoder.stages.{i}.downsampling_layer.1.weight""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.1.bias""", F"""backbone.encoder.stages.{i}.downsampling_layer.1.bias""") ) rename_keys.append((F"""backbone.norm{i}.weight""", F"""backbone.hidden_states_norms.stage{i+1}.weight""") ) rename_keys.append((F"""backbone.norm{i}.bias""", F"""backbone.hidden_states_norms.stage{i+1}.bias""") ) # decode head rename_keys.extend( [ ('decode_head.conv_seg.weight', 'decode_head.classifier.weight'), ('decode_head.conv_seg.bias', 'decode_head.classifier.bias'), ('auxiliary_head.conv_seg.weight', 'auxiliary_head.classifier.weight'), ('auxiliary_head.conv_seg.bias', 'auxiliary_head.classifier.bias'), ] ) # fmt: on return rename_keys def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Optional[int]: _lowercase : Any = dct.pop(SCREAMING_SNAKE_CASE ) _lowercase : Any = val def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Any: _lowercase : List[Any] = { 'upernet-convnext-tiny': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_tiny_fp16_512x512_160k_ade20k/upernet_convnext_tiny_fp16_512x512_160k_ade20k_20220227_124553-cad485de.pth', 'upernet-convnext-small': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_small_fp16_512x512_160k_ade20k/upernet_convnext_small_fp16_512x512_160k_ade20k_20220227_131208-1b1e394f.pth', 'upernet-convnext-base': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_base_fp16_512x512_160k_ade20k/upernet_convnext_base_fp16_512x512_160k_ade20k_20220227_181227-02a24fc6.pth', 'upernet-convnext-large': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_large_fp16_640x640_160k_ade20k/upernet_convnext_large_fp16_640x640_160k_ade20k_20220226_040532-e57aa54d.pth', 'upernet-convnext-xlarge': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_xlarge_fp16_640x640_160k_ade20k/upernet_convnext_xlarge_fp16_640x640_160k_ade20k_20220226_080344-95fc38c2.pth', } _lowercase : Optional[int] = model_name_to_url[model_name] _lowercase : str = torch.hub.load_state_dict_from_url(SCREAMING_SNAKE_CASE , map_location='cpu' )['state_dict'] _lowercase : Optional[int] = get_upernet_config(SCREAMING_SNAKE_CASE ) _lowercase : Tuple = UperNetForSemanticSegmentation(SCREAMING_SNAKE_CASE ) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): _lowercase : List[Any] = state_dict.pop(SCREAMING_SNAKE_CASE ) if "bn" in key: _lowercase : Any = key.replace('bn' , 'batch_norm' ) _lowercase : Any = val # rename keys _lowercase : int = create_rename_keys(SCREAMING_SNAKE_CASE ) for src, dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) model.load_state_dict(SCREAMING_SNAKE_CASE ) # verify on image _lowercase : Union[str, Any] = 'https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg' _lowercase : Any = Image.open(requests.get(SCREAMING_SNAKE_CASE , stream=SCREAMING_SNAKE_CASE ).raw ).convert('RGB' ) _lowercase : Tuple = SegformerImageProcessor() _lowercase : Tuple = processor(SCREAMING_SNAKE_CASE , return_tensors='pt' ).pixel_values with torch.no_grad(): _lowercase : Dict = model(SCREAMING_SNAKE_CASE ) if model_name == "upernet-convnext-tiny": _lowercase : Dict = torch.tensor( [[-8.8110, -8.8110, -8.6521], [-8.8110, -8.8110, -8.6521], [-8.7746, -8.7746, -8.6130]] ) elif model_name == "upernet-convnext-small": _lowercase : Union[str, Any] = torch.tensor( [[-8.8236, -8.8236, -8.6771], [-8.8236, -8.8236, -8.6771], [-8.7638, -8.7638, -8.6240]] ) elif model_name == "upernet-convnext-base": _lowercase : Dict = torch.tensor( [[-8.8558, -8.8558, -8.6905], [-8.8558, -8.8558, -8.6905], [-8.7669, -8.7669, -8.6021]] ) elif model_name == "upernet-convnext-large": _lowercase : Optional[int] = torch.tensor( [[-8.6660, -8.6660, -8.6210], [-8.6660, -8.6660, -8.6210], [-8.6310, -8.6310, -8.5964]] ) elif model_name == "upernet-convnext-xlarge": _lowercase : str = torch.tensor( [[-8.4980, -8.4980, -8.3977], [-8.4980, -8.4980, -8.3977], [-8.4379, -8.4379, -8.3412]] ) print('Logits:' , outputs.logits[0, 0, :3, :3] ) assert torch.allclose(outputs.logits[0, 0, :3, :3] , SCREAMING_SNAKE_CASE , atol=1E-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(SCREAMING_SNAKE_CASE ) print(F"""Saving processor to {pytorch_dump_folder_path}""" ) processor.save_pretrained(SCREAMING_SNAKE_CASE ) if push_to_hub: print(F"""Pushing model and processor for {model_name} to hub""" ) model.push_to_hub(F"""openmmlab/{model_name}""" ) processor.push_to_hub(F"""openmmlab/{model_name}""" ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="upernet-convnext-tiny", type=str, choices=[f'''upernet-convnext-{size}''' for size in ["tiny", "small", "base", "large", "xlarge"]], help="Name of the ConvNext UperNet model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) UpperCamelCase = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
66
1
import argparse from pathlib import Path import fairseq import torch from fairseq.models.xmod import XMODModel as FairseqXmodModel from packaging import version from transformers import XmodConfig, XmodForMaskedLM, XmodForSequenceClassification from transformers.utils import logging if version.parse(fairseq.__version__) < version.parse("0.12.2"): raise Exception("requires fairseq >= 0.12.2") if version.parse(fairseq.__version__) > version.parse("2"): raise Exception("requires fairseq < v2") logging.set_verbosity_info() UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = "Hello, World!" UpperCamelCase = "en_XX" def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Optional[Any]: _lowercase : List[Any] = Path('data_bin' ) _lowercase : str = FairseqXmodModel.from_pretrained( model_name_or_path=str(Path(SCREAMING_SNAKE_CASE ).parent ) , checkpoint_file=Path(SCREAMING_SNAKE_CASE ).name , _name='xmod_base' , arch='xmod_base' , task='multilingual_masked_lm' , data_name_or_path=str(SCREAMING_SNAKE_CASE ) , bpe='sentencepiece' , sentencepiece_model=str(Path(SCREAMING_SNAKE_CASE ).parent / 'sentencepiece.bpe.model' ) , src_dict=str(data_dir / 'dict.txt' ) , ) xmod.eval() # disable dropout print(SCREAMING_SNAKE_CASE ) _lowercase : str = xmod.model.encoder.sentence_encoder _lowercase : List[str] = XmodConfig( vocab_size=xmod_sent_encoder.embed_tokens.num_embeddings , hidden_size=xmod.cfg.model.encoder_embed_dim , num_hidden_layers=xmod.cfg.model.encoder_layers , num_attention_heads=xmod.cfg.model.encoder_attention_heads , intermediate_size=xmod.cfg.model.encoder_ffn_embed_dim , max_position_embeddings=514 , type_vocab_size=1 , layer_norm_eps=1E-5 , pre_norm=xmod.cfg.model.encoder_normalize_before , adapter_reduction_factor=getattr(xmod.cfg.model , 'bottleneck' , 2 ) , adapter_layer_norm=xmod.cfg.model.adapter_layer_norm , adapter_reuse_layer_norm=xmod.cfg.model.adapter_reuse_layer_norm , ln_before_adapter=xmod.cfg.model.ln_before_adapter , languages=xmod.cfg.model.languages , ) if classification_head: _lowercase : List[str] = xmod.model.classification_heads['mnli'].out_proj.weight.shape[0] print('Our X-MOD config:' , SCREAMING_SNAKE_CASE ) _lowercase : Tuple = XmodForSequenceClassification(SCREAMING_SNAKE_CASE ) if classification_head else XmodForMaskedLM(SCREAMING_SNAKE_CASE ) model.eval() # Now let's copy all the weights. # Embeddings _lowercase : int = xmod_sent_encoder.embed_tokens.weight _lowercase : Union[str, Any] = xmod_sent_encoder.embed_positions.weight _lowercase : Union[str, Any] = torch.zeros_like( model.roberta.embeddings.token_type_embeddings.weight ) # just zero them out b/c xmod doesn't use them. _lowercase : str = xmod_sent_encoder.layernorm_embedding.weight _lowercase : List[Any] = xmod_sent_encoder.layernorm_embedding.bias for i in range(config.num_hidden_layers ): # Encoder: start of layer _lowercase : str = model.roberta.encoder.layer[i] _lowercase : Dict = xmod_sent_encoder.layers[i] # self attention _lowercase : List[Any] = layer.attention.self if not ( xmod_layer.self_attn.k_proj.weight.data.shape == xmod_layer.self_attn.q_proj.weight.data.shape == xmod_layer.self_attn.v_proj.weight.data.shape == torch.Size((config.hidden_size, config.hidden_size) ) ): raise AssertionError('Dimensions of self-attention weights do not match.' ) _lowercase : Optional[int] = xmod_layer.self_attn.q_proj.weight _lowercase : Dict = xmod_layer.self_attn.q_proj.bias _lowercase : int = xmod_layer.self_attn.k_proj.weight _lowercase : str = xmod_layer.self_attn.k_proj.bias _lowercase : Dict = xmod_layer.self_attn.v_proj.weight _lowercase : int = xmod_layer.self_attn.v_proj.bias # self-attention output _lowercase : List[str] = layer.attention.output if self_output.dense.weight.shape != xmod_layer.self_attn.out_proj.weight.shape: raise AssertionError('Dimensions of self-attention output weights do not match.' ) _lowercase : Any = xmod_layer.self_attn.out_proj.weight _lowercase : Dict = xmod_layer.self_attn.out_proj.bias _lowercase : Union[str, Any] = xmod_layer.self_attn_layer_norm.weight _lowercase : Optional[int] = xmod_layer.self_attn_layer_norm.bias # intermediate _lowercase : Optional[int] = layer.intermediate if intermediate.dense.weight.shape != xmod_layer.fca.weight.shape: raise AssertionError('Dimensions of intermediate weights do not match.' ) _lowercase : Optional[Any] = xmod_layer.fca.weight _lowercase : Tuple = xmod_layer.fca.bias # output _lowercase : List[str] = layer.output if bert_output.dense.weight.shape != xmod_layer.fca.weight.shape: raise AssertionError('Dimensions of feed-forward weights do not match.' ) _lowercase : Tuple = xmod_layer.fca.weight _lowercase : Optional[Any] = xmod_layer.fca.bias _lowercase : Optional[int] = xmod_layer.final_layer_norm.weight _lowercase : List[Any] = xmod_layer.final_layer_norm.bias if bert_output.adapter_layer_norm is not None: _lowercase : Optional[int] = xmod_layer.adapter_layer_norm.weight _lowercase : int = xmod_layer.adapter_layer_norm.bias if sorted(bert_output.adapter_modules.keys() ) != sorted(xmod_layer.adapter_modules.keys() ): raise AssertionError('Lists of language adapters do not match.' ) for lang_code, adapter in xmod_layer.adapter_modules.items(): _lowercase : Dict = bert_output.adapter_modules[lang_code] _lowercase : Tuple = xmod_layer.adapter_modules[lang_code] _lowercase : int = from_adapter.fca.weight _lowercase : str = from_adapter.fca.bias _lowercase : int = from_adapter.fca.weight _lowercase : Tuple = from_adapter.fca.bias # end of layer if xmod_sent_encoder.layer_norm is not None: _lowercase : int = xmod_sent_encoder.layer_norm.weight _lowercase : List[str] = xmod_sent_encoder.layer_norm.bias if classification_head: _lowercase : Dict = xmod.model.classification_heads['mnli'].dense.weight _lowercase : int = xmod.model.classification_heads['mnli'].dense.bias _lowercase : List[Any] = xmod.model.classification_heads['mnli'].out_proj.weight _lowercase : Optional[int] = xmod.model.classification_heads['mnli'].out_proj.bias else: # LM Head _lowercase : Union[str, Any] = xmod.model.encoder.lm_head.dense.weight _lowercase : str = xmod.model.encoder.lm_head.dense.bias _lowercase : Dict = xmod.model.encoder.lm_head.layer_norm.weight _lowercase : Any = xmod.model.encoder.lm_head.layer_norm.bias _lowercase : List[Any] = xmod.model.encoder.lm_head.weight _lowercase : Dict = xmod.model.encoder.lm_head.bias # Let's check that we get the same results. _lowercase : int = xmod.encode(SCREAMING_SNAKE_CASE ).unsqueeze(0 ) # batch of size 1 model.roberta.set_default_language(SCREAMING_SNAKE_CASE ) _lowercase : Optional[int] = model(SCREAMING_SNAKE_CASE )[0] if classification_head: _lowercase : Union[str, Any] = xmod.model.classification_heads['mnli'](xmod.extract_features(SCREAMING_SNAKE_CASE ) ) else: _lowercase : Tuple = xmod.model(SCREAMING_SNAKE_CASE , lang_id=[SAMPLE_LANGUAGE] )[0] print(our_output.shape , their_output.shape ) _lowercase : Tuple = torch.max(torch.abs(our_output - their_output ) ).item() print(F"""max_absolute_diff = {max_absolute_diff}""" ) # ~ 1e-7 _lowercase : int = torch.allclose(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , atol=1E-3 ) print('Do both models output the same tensors?' , '🔥' if success else '💩' ) if not success: raise Exception('Something went wRoNg' ) Path(SCREAMING_SNAKE_CASE ).mkdir(parents=SCREAMING_SNAKE_CASE , exist_ok=SCREAMING_SNAKE_CASE ) print(F"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--xmod_checkpoint_path", default=None, type=str, required=True, help="Path the official PyTorch dump." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--classification_head", action="store_true", help="Whether to convert a final classification head." ) UpperCamelCase = parser.parse_args() convert_xmod_checkpoint_to_pytorch( args.xmod_checkpoint_path, args.pytorch_dump_folder_path, args.classification_head )
66
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto.configuration_auto import CONFIG_MAPPING UpperCamelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : Any = "upernet" def __init__( self , _lowerCAmelCase=None , _lowerCAmelCase=5_1_2 , _lowerCAmelCase=0.02 , _lowerCAmelCase=[1, 2, 3, 6] , _lowerCAmelCase=True , _lowerCAmelCase=0.4 , _lowerCAmelCase=3_8_4 , _lowerCAmelCase=2_5_6 , _lowerCAmelCase=1 , _lowerCAmelCase=False , _lowerCAmelCase=2_5_5 , **_lowerCAmelCase , ): super().__init__(**_lowerCAmelCase ) if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.' ) _lowercase : Optional[Any] = CONFIG_MAPPING['resnet'](out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): _lowercase : List[Any] = backbone_config.get('model_type' ) _lowercase : str = CONFIG_MAPPING[backbone_model_type] _lowercase : Tuple = config_class.from_dict(_lowerCAmelCase ) _lowercase : Optional[Any] = backbone_config _lowercase : Any = hidden_size _lowercase : Any = initializer_range _lowercase : Tuple = pool_scales _lowercase : List[Any] = use_auxiliary_head _lowercase : Optional[Any] = auxiliary_loss_weight _lowercase : Any = auxiliary_in_channels _lowercase : Any = auxiliary_channels _lowercase : List[str] = auxiliary_num_convs _lowercase : List[str] = auxiliary_concat_input _lowercase : Tuple = loss_ignore_index def __a ( self ): _lowercase : str = copy.deepcopy(self.__dict__ ) _lowercase : Tuple = self.backbone_config.to_dict() _lowercase : int = self.__class__.model_type return output
66
1
import gc import random import unittest import numpy as np import torch from diffusers import DDIMScheduler, KandinskyVaaPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : List[Any] = KandinskyVaaPipeline _UpperCamelCase : int = [ "image_embeds", "negative_image_embeds", ] _UpperCamelCase : Optional[int] = ["image_embeds", "negative_image_embeds"] _UpperCamelCase : Optional[Any] = [ "generator", "height", "width", "latents", "guidance_scale", "num_inference_steps", "return_dict", "guidance_scale", "num_images_per_prompt", "output_type", "return_dict", ] _UpperCamelCase : Dict = False @property def __a ( self ): return 3_2 @property def __a ( self ): return 3_2 @property def __a ( self ): return self.time_input_dim @property def __a ( self ): return self.time_input_dim * 4 @property def __a ( self ): return 1_0_0 @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Tuple = { 'in_channels': 4, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'image', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } _lowercase : Union[str, Any] = UNetaDConditionModel(**_lowerCAmelCase ) return model @property def __a ( self ): return { "block_out_channels": [3_2, 6_4], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 1_2, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Union[str, Any] = VQModel(**self.dummy_movq_kwargs ) return model def __a ( self ): _lowercase : List[str] = self.dummy_unet _lowercase : Optional[int] = self.dummy_movq _lowercase : Tuple = DDIMScheduler( num_train_timesteps=1_0_0_0 , beta_schedule='linear' , beta_start=0.0_00_85 , beta_end=0.0_12 , clip_sample=_lowerCAmelCase , set_alpha_to_one=_lowerCAmelCase , steps_offset=1 , prediction_type='epsilon' , thresholding=_lowerCAmelCase , ) _lowercase : str = { 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): _lowercase : Optional[int] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : Tuple = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( _lowerCAmelCase ) if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : Any = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Tuple = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : Optional[int] = { 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'generator': generator, 'height': 6_4, 'width': 6_4, 'guidance_scale': 4.0, 'num_inference_steps': 2, 'output_type': 'np', } return inputs def __a ( self ): _lowercase : Union[str, Any] = 'cpu' _lowercase : Any = self.get_dummy_components() _lowercase : Any = self.pipeline_class(**_lowerCAmelCase ) _lowercase : str = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Tuple = pipe(**self.get_dummy_inputs(_lowerCAmelCase ) ) _lowercase : Dict = output.images _lowercase : str = pipe( **self.get_dummy_inputs(_lowerCAmelCase ) , return_dict=_lowerCAmelCase , )[0] _lowercase : str = image[0, -3:, -3:, -1] _lowercase : List[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) _lowercase : Any = np.array( [0.6_23_79_76, 1.0, 0.36_44_13_32, 1.0, 0.70_63_96_34, 0.29_87_71_86, 0.85_65_21_25, 0.5_21_68_43, 0.54_45_40_46] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), F""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), F""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" @slow @require_torch_gpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ): _lowercase : List[str] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinskyv22/kandinskyv22_text2img_cat_fp16.npy' ) _lowercase : Optional[Any] = KandinskyVaaPriorPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-prior' , torch_dtype=torch.floataa ) pipe_prior.to(_lowerCAmelCase ) _lowercase : int = KandinskyVaaPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-decoder' , torch_dtype=torch.floataa ) _lowercase : List[Any] = pipeline.to(_lowerCAmelCase ) pipeline.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : List[Any] = 'red cat, 4k photo' _lowercase : List[str] = torch.Generator(device='cuda' ).manual_seed(0 ) _lowercase , _lowercase : Union[str, Any] = pipe_prior( _lowerCAmelCase , generator=_lowerCAmelCase , num_inference_steps=5 , negative_prompt='' , ).to_tuple() _lowercase : Any = torch.Generator(device='cuda' ).manual_seed(0 ) _lowercase : Union[str, Any] = pipeline( image_embeds=_lowerCAmelCase , negative_image_embeds=_lowerCAmelCase , generator=_lowerCAmelCase , num_inference_steps=1_0_0 , output_type='np' , ) _lowercase : Optional[int] = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert_mean_pixel_difference(_lowerCAmelCase , _lowerCAmelCase )
66
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: if a < 0 or b < 0: raise ValueError('the value of both inputs must be positive' ) _lowercase : str = str(bin(SCREAMING_SNAKE_CASE ) )[2:] # remove the leading "0b" _lowercase : int = str(bin(SCREAMING_SNAKE_CASE ) )[2:] # remove the leading "0b" _lowercase : List[Any] = max(len(SCREAMING_SNAKE_CASE ) , len(SCREAMING_SNAKE_CASE ) ) return "0b" + "".join( str(int(char_a == '1' and char_b == '1' ) ) for char_a, char_b in zip(a_binary.zfill(SCREAMING_SNAKE_CASE ) , b_binary.zfill(SCREAMING_SNAKE_CASE ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
66
1
import argparse import json import numpy import torch from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> int: # Load checkpoint _lowercase : Optional[int] = torch.load(SCREAMING_SNAKE_CASE , map_location='cpu' ) _lowercase : Tuple = chkpt['model'] # We have the base model one level deeper than the original XLM repository _lowercase : Optional[int] = {} for k, v in state_dict.items(): if "pred_layer" in k: _lowercase : Optional[Any] = v else: _lowercase : Optional[Any] = v _lowercase : Optional[int] = chkpt['params'] _lowercase : Optional[Any] = {n: v for n, v in config.items() if not isinstance(SCREAMING_SNAKE_CASE , (torch.FloatTensor, numpy.ndarray) )} _lowercase : Optional[int] = chkpt['dico_word2id'] _lowercase : Tuple = {s + '</w>' if s.find('@@' ) == -1 and i > 13 else s.replace('@@' , '' ): i for s, i in vocab.items()} # Save pytorch-model _lowercase : str = pytorch_dump_folder_path + '/' + WEIGHTS_NAME _lowercase : int = pytorch_dump_folder_path + '/' + CONFIG_NAME _lowercase : Optional[Any] = pytorch_dump_folder_path + '/' + VOCAB_FILES_NAMES['vocab_file'] print(F"""Save PyTorch model to {pytorch_weights_dump_path}""" ) torch.save(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) print(F"""Save configuration file to {pytorch_config_dump_path}""" ) with open(SCREAMING_SNAKE_CASE , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(SCREAMING_SNAKE_CASE , indent=2 ) + '\n' ) print(F"""Save vocab file to {pytorch_config_dump_path}""" ) with open(SCREAMING_SNAKE_CASE , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(SCREAMING_SNAKE_CASE , indent=2 ) + '\n' ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--xlm_checkpoint_path", default=None, type=str, required=True, help="Path the official PyTorch dump." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) UpperCamelCase = parser.parse_args() convert_xlm_checkpoint_to_pytorch(args.xlm_checkpoint_path, args.pytorch_dump_folder_path)
66
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class lowerCAmelCase_ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCamelCase : int = IFInpaintingSuperResolutionPipeline _UpperCamelCase : Any = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"width", "height"} _UpperCamelCase : Union[str, Any] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"original_image"} ) _UpperCamelCase : Tuple = PipelineTesterMixin.required_optional_params - {"latents"} def __a ( self ): return self._get_superresolution_dummy_components() def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : int = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Union[str, Any] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : Union[str, Any] = floats_tensor((1, 3, 1_6, 1_6) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : Optional[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'original_image': original_image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def __a ( self ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def __a ( self ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != 'cuda' , reason='float16 requires CUDA' ) def __a ( self ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __a ( self ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __a ( self ): self._test_save_load_local() def __a ( self ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
66
1
import sys UpperCamelCase = ( "73167176531330624919225119674426574742355349194934" "96983520312774506326239578318016984801869478851843" "85861560789112949495459501737958331952853208805511" "12540698747158523863050715693290963295227443043557" "66896648950445244523161731856403098711121722383113" "62229893423380308135336276614282806444486645238749" "30358907296290491560440772390713810515859307960866" "70172427121883998797908792274921901699720888093776" "65727333001053367881220235421809751254540594752243" "52584907711670556013604839586446706324415722155397" "53697817977846174064955149290862569321978468622482" "83972241375657056057490261407972968652414535100474" "82166370484403199890008895243450658541227588666881" "16427171479924442928230863465674813919123162824586" "17866458359124566529476545682848912883142607690042" "24219022671055626321111109370544217506941658960408" "07198403850962455444362981230987879927244284909188" "84580156166097919133875499200524063689912560717606" "05886116467109405077541002256983155200055935729725" "71636269561882670428252483600823257530420752963450" ) def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : List[Any] = 1 for digit in s: product *= int(SCREAMING_SNAKE_CASE ) return product def __magic_name__ ( SCREAMING_SNAKE_CASE = N ) -> int: _lowercase : Dict = -sys.maxsize - 1 _lowercase : Tuple = n[:13] _lowercase : List[Any] = 13 while cur_index < len(SCREAMING_SNAKE_CASE ) - 13: if int(n[cur_index] ) >= int(substr[0] ): _lowercase : List[str] = substr[1:] + n[cur_index] cur_index += 1 else: _lowercase : str = max(SCREAMING_SNAKE_CASE , str_eval(SCREAMING_SNAKE_CASE ) ) _lowercase : Dict = n[cur_index : cur_index + 13] cur_index += 13 return largest_product if __name__ == "__main__": print(f'''{solution() = }''')
66
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> tuple[float, float]: # Check if the input is valid if not len(SCREAMING_SNAKE_CASE ) == len(SCREAMING_SNAKE_CASE ) == 3: raise ValueError('Please enter a valid equation.' ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError('Both a & b of two equations can\'t be zero.' ) # Extract the coefficients _lowercase , _lowercase , _lowercase : Tuple = equationa _lowercase , _lowercase , _lowercase : Dict = equationa # Calculate the determinants of the matrices _lowercase : str = aa * ba - aa * ba _lowercase : Any = ca * ba - ca * ba _lowercase : Optional[int] = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError('Infinite solutions. (Consistent system)' ) else: raise ValueError('No solution. (Inconsistent system)' ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: _lowercase : Union[str, Any] = determinant_x / determinant _lowercase : Tuple = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
66
1
from __future__ import annotations UpperCamelCase = tuple[int, int, int] UpperCamelCase = tuple[str, str, str] # used alphabet -------------------------- # from string.ascii_uppercase UpperCamelCase = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" # -------------------------- default selection -------------------------- # rotors -------------------------- UpperCamelCase = "EGZWVONAHDCLFQMSIPJBYUKXTR" UpperCamelCase = "FOBHMDKEXQNRAULPGSJVTYICZW" UpperCamelCase = "ZJXESIUQLHAVRMDOYGTNFWPBKC" # reflector -------------------------- UpperCamelCase = { "A": "N", "N": "A", "B": "O", "O": "B", "C": "P", "P": "C", "D": "Q", "Q": "D", "E": "R", "R": "E", "F": "S", "S": "F", "G": "T", "T": "G", "H": "U", "U": "H", "I": "V", "V": "I", "J": "W", "W": "J", "K": "X", "X": "K", "L": "Y", "Y": "L", "M": "Z", "Z": "M", } # -------------------------- extra rotors -------------------------- UpperCamelCase = "RMDJXFUWGISLHVTCQNKYPBEZOA" UpperCamelCase = "SGLCPQWZHKXAREONTFBVIYJUDM" UpperCamelCase = "HVSICLTYKQUBXDWAJZOMFGPREN" UpperCamelCase = "RZWQHFMVDBKICJLNTUXAGYPSOE" UpperCamelCase = "LFKIJODBEGAMQPXVUHYSTCZRWN" UpperCamelCase = "KOAEGVDHXPQZMLFTYWJNBRCIUS" def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> tuple[RotorPositionT, RotorSelectionT, dict[str, str]]: # Checks if there are 3 unique rotors if (unique_rotsel := len(set(SCREAMING_SNAKE_CASE ) )) < 3: _lowercase : Optional[int] = F"""Please use 3 unique rotors (not {unique_rotsel})""" raise Exception(SCREAMING_SNAKE_CASE ) # Checks if rotor positions are valid _lowercase , _lowercase , _lowercase : int = rotpos if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : Dict = F"""First rotor position is not within range of 1..26 ({rotorposa}""" raise ValueError(SCREAMING_SNAKE_CASE ) if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : int = F"""Second rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(SCREAMING_SNAKE_CASE ) if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : str = F"""Third rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(SCREAMING_SNAKE_CASE ) # Validates string and returns dict _lowercase : Tuple = _plugboard(SCREAMING_SNAKE_CASE ) return rotpos, rotsel, pbdict def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> dict[str, str]: # tests the input string if it # a) is type string # b) has even length (so pairs can be made) if not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): _lowercase : Optional[int] = F"""Plugboard setting isn't type string ({type(SCREAMING_SNAKE_CASE )})""" raise TypeError(SCREAMING_SNAKE_CASE ) elif len(SCREAMING_SNAKE_CASE ) % 2 != 0: _lowercase : Optional[int] = F"""Odd number of symbols ({len(SCREAMING_SNAKE_CASE )})""" raise Exception(SCREAMING_SNAKE_CASE ) elif pbstring == "": return {} pbstring.replace(' ' , '' ) # Checks if all characters are unique _lowercase : Dict = set() for i in pbstring: if i not in abc: _lowercase : str = F"""'{i}' not in list of symbols""" raise Exception(SCREAMING_SNAKE_CASE ) elif i in tmppbl: _lowercase : int = F"""Duplicate symbol ({i})""" raise Exception(SCREAMING_SNAKE_CASE ) else: tmppbl.add(SCREAMING_SNAKE_CASE ) del tmppbl # Created the dictionary _lowercase : Optional[Any] = {} for j in range(0 , len(SCREAMING_SNAKE_CASE ) - 1 , 2 ): _lowercase : Dict = pbstring[j + 1] _lowercase : Union[str, Any] = pbstring[j] return pb def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = (rotora, rotora, rotora) , SCREAMING_SNAKE_CASE = "" , ) -> str: _lowercase : List[str] = text.upper() _lowercase , _lowercase , _lowercase : List[str] = _validator( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , plugb.upper() ) _lowercase , _lowercase , _lowercase : Optional[int] = rotor_position _lowercase , _lowercase , _lowercase : Union[str, Any] = rotor_selection rotorposa -= 1 rotorposa -= 1 rotorposa -= 1 _lowercase : Optional[int] = [] # encryption/decryption process -------------------------- for symbol in text: if symbol in abc: # 1st plugboard -------------------------- if symbol in plugboard: _lowercase : Dict = plugboard[symbol] # rotor ra -------------------------- _lowercase : Optional[Any] = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : Union[str, Any] = rotora[index % len(SCREAMING_SNAKE_CASE )] # rotor rb -------------------------- _lowercase : Tuple = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : str = rotora[index % len(SCREAMING_SNAKE_CASE )] # rotor rc -------------------------- _lowercase : List[Any] = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : List[str] = rotora[index % len(SCREAMING_SNAKE_CASE )] # reflector -------------------------- # this is the reason you don't need another machine to decipher _lowercase : List[str] = reflector[symbol] # 2nd rotors _lowercase : List[str] = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] _lowercase : Tuple = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] _lowercase : Dict = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] # 2nd plugboard if symbol in plugboard: _lowercase : int = plugboard[symbol] # moves/resets rotor positions rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : Any = 0 rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : int = 0 rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : Any = 0 # else: # pass # Error could be also raised # raise ValueError( # 'Invalid symbol('+repr(symbol)+')') result.append(SCREAMING_SNAKE_CASE ) return "".join(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": UpperCamelCase = "This is my Python script that emulates the Enigma machine from WWII." UpperCamelCase = (1, 1, 1) UpperCamelCase = "pictures" UpperCamelCase = (rotora, rotora, rotora) UpperCamelCase = enigma(message, rotor_pos, rotor_sel, pb) print("Encrypted message:", en) print("Decrypted message:", enigma(en, rotor_pos, rotor_sel, pb))
66
def __magic_name__ ( SCREAMING_SNAKE_CASE = 50 ) -> int: _lowercase : Optional[int] = [[0] * 3 for _ in range(length + 1 )] for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): different_colour_ways_number[row_length][tile_length - 2] += ( different_colour_ways_number[row_length - tile_start - tile_length][ tile_length - 2 ] + 1 ) return sum(different_colour_ways_number[length] ) if __name__ == "__main__": print(f'''{solution() = }''')
66
1
import gc import unittest import numpy as np import torch from diffusers import StableDiffusionKDiffusionPipeline from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() @slow @require_torch_gpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ): _lowercase : int = StableDiffusionKDiffusionPipeline.from_pretrained('CompVis/stable-diffusion-v1-4' ) _lowercase : Union[str, Any] = sd_pipe.to(_lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=_lowerCAmelCase ) sd_pipe.set_scheduler('sample_euler' ) _lowercase : Any = 'A painting of a squirrel eating a burger' _lowercase : Any = torch.manual_seed(0 ) _lowercase : Union[str, Any] = sd_pipe([prompt] , generator=_lowerCAmelCase , guidance_scale=9.0 , num_inference_steps=2_0 , output_type='np' ) _lowercase : Tuple = output.images _lowercase : List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) _lowercase : Any = np.array([0.04_47, 0.04_92, 0.04_68, 0.04_08, 0.03_83, 0.04_08, 0.03_54, 0.03_80, 0.03_39] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __a ( self ): _lowercase : List[Any] = StableDiffusionKDiffusionPipeline.from_pretrained('stabilityai/stable-diffusion-2-1-base' ) _lowercase : List[str] = sd_pipe.to(_lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=_lowerCAmelCase ) sd_pipe.set_scheduler('sample_euler' ) _lowercase : Optional[Any] = 'A painting of a squirrel eating a burger' _lowercase : List[str] = torch.manual_seed(0 ) _lowercase : str = sd_pipe([prompt] , generator=_lowerCAmelCase , guidance_scale=9.0 , num_inference_steps=2_0 , output_type='np' ) _lowercase : List[Any] = output.images _lowercase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) _lowercase : Union[str, Any] = np.array([0.12_37, 0.13_20, 0.14_38, 0.13_59, 0.13_90, 0.11_32, 0.12_77, 0.11_75, 0.11_12] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-1 def __a ( self ): _lowercase : int = StableDiffusionKDiffusionPipeline.from_pretrained('stabilityai/stable-diffusion-2-1-base' ) _lowercase : Any = sd_pipe.to(_lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=_lowerCAmelCase ) sd_pipe.set_scheduler('sample_dpmpp_2m' ) _lowercase : int = 'A painting of a squirrel eating a burger' _lowercase : Tuple = torch.manual_seed(0 ) _lowercase : Optional[Any] = sd_pipe( [prompt] , generator=_lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=1_5 , output_type='np' , use_karras_sigmas=_lowerCAmelCase , ) _lowercase : Any = output.images _lowercase : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) _lowercase : List[Any] = np.array( [0.11_38_16_89, 0.12_11_29_21, 0.1_38_94_57, 0.12_54_96_06, 0.1_24_49_64, 0.10_83_15_17, 0.11_56_28_66, 0.10_86_78_16, 0.10_49_90_48] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
66
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) UpperCamelCase = { "configuration_convnext": ["CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP", "ConvNextConfig", "ConvNextOnnxConfig"] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["ConvNextFeatureExtractor"] UpperCamelCase = ["ConvNextImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST", "ConvNextForImageClassification", "ConvNextModel", "ConvNextPreTrainedModel", "ConvNextBackbone", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "TFConvNextForImageClassification", "TFConvNextModel", "TFConvNextPreTrainedModel", ] if TYPE_CHECKING: from .configuration_convnext import CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvNextConfig, ConvNextOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_convnext import ConvNextFeatureExtractor from .image_processing_convnext import ConvNextImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convnext import ( CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvNextBackbone, ConvNextForImageClassification, ConvNextModel, ConvNextPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convnext import TFConvNextForImageClassification, TFConvNextModel, TFConvNextPreTrainedModel else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure)
66
1
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetrImageProcessor class lowerCAmelCase_ ( unittest.TestCase ): def __init__( self , _lowerCAmelCase , _lowerCAmelCase=7 , _lowerCAmelCase=3 , _lowerCAmelCase=3_0 , _lowerCAmelCase=4_0_0 , _lowerCAmelCase=True , _lowerCAmelCase=None , _lowerCAmelCase=True , _lowerCAmelCase=1 / 2_5_5 , _lowerCAmelCase=True , _lowerCAmelCase=[0.5, 0.5, 0.5] , _lowerCAmelCase=[0.5, 0.5, 0.5] , _lowerCAmelCase=True , ): # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p _lowercase : Dict = size if size is not None else {'shortest_edge': 1_8, 'longest_edge': 1_3_3_3} _lowercase : Optional[Any] = parent _lowercase : Optional[int] = batch_size _lowercase : int = num_channels _lowercase : Optional[int] = min_resolution _lowercase : Union[str, Any] = max_resolution _lowercase : List[Any] = do_resize _lowercase : Dict = size _lowercase : Dict = do_rescale _lowercase : Dict = rescale_factor _lowercase : Union[str, Any] = do_normalize _lowercase : int = image_mean _lowercase : str = image_std _lowercase : List[Any] = do_pad def __a ( self ): return { "do_resize": self.do_resize, "size": self.size, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_pad": self.do_pad, } def __a ( self , _lowerCAmelCase , _lowerCAmelCase=False ): if not batched: _lowercase : Optional[int] = image_inputs[0] if isinstance(_lowerCAmelCase , Image.Image ): _lowercase , _lowercase : Dict = image.size else: _lowercase , _lowercase : Optional[Any] = image.shape[1], image.shape[2] if w < h: _lowercase : Optional[Any] = int(self.size['shortest_edge'] * h / w ) _lowercase : Tuple = self.size['shortest_edge'] elif w > h: _lowercase : List[str] = self.size['shortest_edge'] _lowercase : Any = int(self.size['shortest_edge'] * w / h ) else: _lowercase : Optional[int] = self.size['shortest_edge'] _lowercase : Union[str, Any] = self.size['shortest_edge'] else: _lowercase : int = [] for image in image_inputs: _lowercase , _lowercase : Optional[int] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) _lowercase : List[Any] = max(_lowerCAmelCase , key=lambda _lowerCAmelCase : item[0] )[0] _lowercase : List[str] = max(_lowerCAmelCase , key=lambda _lowerCAmelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : List[str] = DetrImageProcessor if is_vision_available() else None def __a ( self ): _lowercase : List[str] = DetrImageProcessingTester(self ) @property def __a ( self ): return self.image_processor_tester.prepare_image_processor_dict() def __a ( self ): _lowercase : Optional[int] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_lowerCAmelCase , 'image_mean' ) ) self.assertTrue(hasattr(_lowerCAmelCase , 'image_std' ) ) self.assertTrue(hasattr(_lowerCAmelCase , 'do_normalize' ) ) self.assertTrue(hasattr(_lowerCAmelCase , 'do_rescale' ) ) self.assertTrue(hasattr(_lowerCAmelCase , 'rescale_factor' ) ) self.assertTrue(hasattr(_lowerCAmelCase , 'do_resize' ) ) self.assertTrue(hasattr(_lowerCAmelCase , 'size' ) ) self.assertTrue(hasattr(_lowerCAmelCase , 'do_pad' ) ) def __a ( self ): _lowercase : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 1_8, 'longest_edge': 1_3_3_3} ) self.assertEqual(image_processor.do_pad , _lowerCAmelCase ) _lowercase : Optional[Any] = self.image_processing_class.from_dict( self.image_processor_dict , size=4_2 , max_size=8_4 , pad_and_return_pixel_mask=_lowerCAmelCase ) self.assertEqual(image_processor.size , {'shortest_edge': 4_2, 'longest_edge': 8_4} ) self.assertEqual(image_processor.do_pad , _lowerCAmelCase ) def __a ( self ): pass def __a ( self ): # Initialize image_processing _lowercase : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _lowercase : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(_lowerCAmelCase , Image.Image ) # Test not batched input _lowercase : List[Any] = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values _lowercase , _lowercase : Tuple = self.image_processor_tester.get_expected_values(_lowerCAmelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowercase , _lowercase : Optional[int] = self.image_processor_tester.get_expected_values(_lowerCAmelCase , batched=_lowerCAmelCase ) _lowercase : int = image_processing(_lowerCAmelCase , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __a ( self ): # Initialize image_processing _lowercase : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _lowercase : Tuple = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCAmelCase , numpify=_lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(_lowerCAmelCase , np.ndarray ) # Test not batched input _lowercase : Optional[int] = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values _lowercase , _lowercase : str = self.image_processor_tester.get_expected_values(_lowerCAmelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowercase : Optional[Any] = image_processing(_lowerCAmelCase , return_tensors='pt' ).pixel_values _lowercase , _lowercase : Optional[Any] = self.image_processor_tester.get_expected_values(_lowerCAmelCase , batched=_lowerCAmelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __a ( self ): # Initialize image_processing _lowercase : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _lowercase : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=_lowerCAmelCase , torchify=_lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(_lowerCAmelCase , torch.Tensor ) # Test not batched input _lowercase : Any = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values _lowercase , _lowercase : str = self.image_processor_tester.get_expected_values(_lowerCAmelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowercase : Dict = image_processing(_lowerCAmelCase , return_tensors='pt' ).pixel_values _lowercase , _lowercase : int = self.image_processor_tester.get_expected_values(_lowerCAmelCase , batched=_lowerCAmelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def __a ( self ): # prepare image and target _lowercase : Optional[Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_annotations.txt' , 'r' ) as f: _lowercase : Tuple = json.loads(f.read() ) _lowercase : List[str] = {'image_id': 3_9_7_6_9, 'annotations': target} # encode them _lowercase : List[Any] = DetrImageProcessor.from_pretrained('facebook/detr-resnet-50' ) _lowercase : Tuple = image_processing(images=_lowerCAmelCase , annotations=_lowerCAmelCase , return_tensors='pt' ) # verify pixel values _lowercase : Tuple = torch.Size([1, 3, 8_0_0, 1_0_6_6] ) self.assertEqual(encoding['pixel_values'].shape , _lowerCAmelCase ) _lowercase : Optional[int] = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , _lowerCAmelCase , atol=1E-4 ) ) # verify area _lowercase : Optional[Any] = torch.tensor([58_87.96_00, 1_12_50.20_61, 48_93_53.84_38, 83_71_22.75_00, 14_79_67.51_56, 16_57_32.34_38] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , _lowerCAmelCase ) ) # verify boxes _lowercase : Dict = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , _lowerCAmelCase ) _lowercase : Optional[Any] = torch.tensor([0.55_03, 0.27_65, 0.06_04, 0.22_15] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , _lowerCAmelCase , atol=1E-3 ) ) # verify image_id _lowercase : List[Any] = torch.tensor([3_9_7_6_9] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , _lowerCAmelCase ) ) # verify is_crowd _lowercase : List[Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , _lowerCAmelCase ) ) # verify class_labels _lowercase : int = torch.tensor([7_5, 7_5, 6_3, 6_5, 1_7, 1_7] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , _lowerCAmelCase ) ) # verify orig_size _lowercase : List[str] = torch.tensor([4_8_0, 6_4_0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , _lowerCAmelCase ) ) # verify size _lowercase : str = torch.tensor([8_0_0, 1_0_6_6] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , _lowerCAmelCase ) ) @slow def __a ( self ): # prepare image, target and masks_path _lowercase : List[str] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt' , 'r' ) as f: _lowercase : int = json.loads(f.read() ) _lowercase : List[str] = {'file_name': '000000039769.png', 'image_id': 3_9_7_6_9, 'segments_info': target} _lowercase : Optional[Any] = pathlib.Path('./tests/fixtures/tests_samples/COCO/coco_panoptic' ) # encode them _lowercase : str = DetrImageProcessor.from_pretrained('facebook/detr-resnet-50-panoptic' ) _lowercase : List[Any] = image_processing(images=_lowerCAmelCase , annotations=_lowerCAmelCase , masks_path=_lowerCAmelCase , return_tensors='pt' ) # verify pixel values _lowercase : str = torch.Size([1, 3, 8_0_0, 1_0_6_6] ) self.assertEqual(encoding['pixel_values'].shape , _lowerCAmelCase ) _lowercase : Dict = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , _lowerCAmelCase , atol=1E-4 ) ) # verify area _lowercase : Dict = torch.tensor([14_79_79.68_75, 16_55_27.04_69, 48_46_38.59_38, 1_12_92.93_75, 58_79.65_62, 76_34.11_47] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , _lowerCAmelCase ) ) # verify boxes _lowercase : int = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , _lowerCAmelCase ) _lowercase : List[Any] = torch.tensor([0.26_25, 0.54_37, 0.46_88, 0.86_25] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , _lowerCAmelCase , atol=1E-3 ) ) # verify image_id _lowercase : List[str] = torch.tensor([3_9_7_6_9] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , _lowerCAmelCase ) ) # verify is_crowd _lowercase : Tuple = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , _lowerCAmelCase ) ) # verify class_labels _lowercase : Union[str, Any] = torch.tensor([1_7, 1_7, 6_3, 7_5, 7_5, 9_3] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , _lowerCAmelCase ) ) # verify masks _lowercase : Optional[int] = 8_2_2_8_7_3 self.assertEqual(encoding['labels'][0]['masks'].sum().item() , _lowerCAmelCase ) # verify orig_size _lowercase : Any = torch.tensor([4_8_0, 6_4_0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , _lowerCAmelCase ) ) # verify size _lowercase : Any = torch.tensor([8_0_0, 1_0_6_6] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , _lowerCAmelCase ) )
66
def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> bool: if p < 2: raise ValueError('p should not be less than 2!' ) elif p == 2: return True _lowercase : Optional[Any] = 4 _lowercase : Tuple = (1 << p) - 1 for _ in range(p - 2 ): _lowercase : Union[str, Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
66
1
import warnings from ...utils import logging from .image_processing_clip import CLIPImageProcessor UpperCamelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( __snake_case ): def __init__( self , *_lowerCAmelCase , **_lowerCAmelCase ): warnings.warn( 'The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use CLIPImageProcessor instead.' , _lowerCAmelCase , ) super().__init__(*_lowerCAmelCase , **_lowerCAmelCase )
66
import random def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = False ) -> dict: _lowercase : dict = {i: [] for i in range(SCREAMING_SNAKE_CASE )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(SCREAMING_SNAKE_CASE ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(SCREAMING_SNAKE_CASE ): for j in range(i + 1 , SCREAMING_SNAKE_CASE ): if random.random() < probability: graph[i].append(SCREAMING_SNAKE_CASE ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(SCREAMING_SNAKE_CASE ) return graph def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> dict: return { i: [j for j in range(SCREAMING_SNAKE_CASE ) if i != j] for i in range(SCREAMING_SNAKE_CASE ) } if __name__ == "__main__": import doctest doctest.testmod()
66
1
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import CLIPImageProcessor, CLIPProcessor @require_vision class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): _lowercase : str = tempfile.mkdtemp() # fmt: off _lowercase : str = ['l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'lo', 'l</w>', 'w</w>', 'r</w>', 't</w>', 'low</w>', 'er</w>', 'lowest</w>', 'newer</w>', 'wider', '<unk>', '<|startoftext|>', '<|endoftext|>'] # fmt: on _lowercase : Any = dict(zip(_lowerCAmelCase , range(len(_lowerCAmelCase ) ) ) ) _lowercase : Any = ['#version: 0.2', 'l o', 'lo w</w>', 'e r</w>', ''] _lowercase : List[str] = {'unk_token': '<unk>'} _lowercase : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) _lowercase : Optional[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(_lowerCAmelCase ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(_lowerCAmelCase ) ) _lowercase : int = { 'do_resize': True, 'size': 2_0, 'do_center_crop': True, 'crop_size': 1_8, 'do_normalize': True, 'image_mean': [0.48_14_54_66, 0.4_57_82_75, 0.40_82_10_73], 'image_std': [0.26_86_29_54, 0.26_13_02_58, 0.27_57_77_11], } _lowercase : str = os.path.join(self.tmpdirname , _lowerCAmelCase ) with open(self.image_processor_file , 'w' , encoding='utf-8' ) as fp: json.dump(_lowerCAmelCase , _lowerCAmelCase ) def __a ( self , **_lowerCAmelCase ): return CLIPTokenizer.from_pretrained(self.tmpdirname , **_lowerCAmelCase ) def __a ( self , **_lowerCAmelCase ): return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **_lowerCAmelCase ) def __a ( self , **_lowerCAmelCase ): return CLIPImageProcessor.from_pretrained(self.tmpdirname , **_lowerCAmelCase ) def __a ( self ): shutil.rmtree(self.tmpdirname ) def __a ( self ): _lowercase : Dict = [np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta )] _lowercase : int = [Image.fromarray(np.moveaxis(_lowerCAmelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __a ( self ): _lowercase : Dict = self.get_tokenizer() _lowercase : Any = self.get_rust_tokenizer() _lowercase : List[str] = self.get_image_processor() _lowercase : List[Any] = CLIPProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) processor_slow.save_pretrained(self.tmpdirname ) _lowercase : Union[str, Any] = CLIPProcessor.from_pretrained(self.tmpdirname , use_fast=_lowerCAmelCase ) _lowercase : Dict = CLIPProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) processor_fast.save_pretrained(self.tmpdirname ) _lowercase : int = CLIPProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , _lowerCAmelCase ) self.assertIsInstance(processor_fast.tokenizer , _lowerCAmelCase ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , _lowerCAmelCase ) self.assertIsInstance(processor_fast.image_processor , _lowerCAmelCase ) def __a ( self ): _lowercase : Union[str, Any] = CLIPProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _lowercase : Optional[Any] = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) _lowercase : List[str] = self.get_image_processor(do_normalize=_lowerCAmelCase , padding_value=1.0 ) _lowercase : List[Any] = CLIPProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=_lowerCAmelCase , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , _lowerCAmelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , _lowerCAmelCase ) def __a ( self ): _lowercase : Dict = self.get_image_processor() _lowercase : Union[str, Any] = self.get_tokenizer() _lowercase : Any = CLIPProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) _lowercase : Any = self.prepare_image_inputs() _lowercase : int = image_processor(_lowerCAmelCase , return_tensors='np' ) _lowercase : List[Any] = processor(images=_lowerCAmelCase , return_tensors='np' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1E-2 ) def __a ( self ): _lowercase : Any = self.get_image_processor() _lowercase : Any = self.get_tokenizer() _lowercase : str = CLIPProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) _lowercase : Dict = 'lower newer' _lowercase : Any = processor(text=_lowerCAmelCase ) _lowercase : Tuple = tokenizer(_lowerCAmelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def __a ( self ): _lowercase : int = self.get_image_processor() _lowercase : Union[str, Any] = self.get_tokenizer() _lowercase : int = CLIPProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) _lowercase : int = 'lower newer' _lowercase : Optional[Any] = self.prepare_image_inputs() _lowercase : List[str] = processor(text=_lowerCAmelCase , images=_lowerCAmelCase ) self.assertListEqual(list(inputs.keys() ) , ['input_ids', 'attention_mask', 'pixel_values'] ) # test if it raises when no input is passed with pytest.raises(_lowerCAmelCase ): processor() def __a ( self ): _lowercase : Optional[Any] = self.get_image_processor() _lowercase : Tuple = self.get_tokenizer() _lowercase : List[Any] = CLIPProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) _lowercase : Dict = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _lowercase : int = processor.batch_decode(_lowerCAmelCase ) _lowercase : Dict = tokenizer.batch_decode(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) def __a ( self ): _lowercase : List[str] = self.get_image_processor() _lowercase : Optional[Any] = self.get_tokenizer() _lowercase : Any = CLIPProcessor(tokenizer=_lowerCAmelCase , image_processor=_lowerCAmelCase ) _lowercase : List[Any] = 'lower newer' _lowercase : List[str] = self.prepare_image_inputs() _lowercase : Dict = processor(text=_lowerCAmelCase , images=_lowerCAmelCase ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
66
from __future__ import annotations UpperCamelCase = tuple[int, int, int] UpperCamelCase = tuple[str, str, str] # used alphabet -------------------------- # from string.ascii_uppercase UpperCamelCase = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" # -------------------------- default selection -------------------------- # rotors -------------------------- UpperCamelCase = "EGZWVONAHDCLFQMSIPJBYUKXTR" UpperCamelCase = "FOBHMDKEXQNRAULPGSJVTYICZW" UpperCamelCase = "ZJXESIUQLHAVRMDOYGTNFWPBKC" # reflector -------------------------- UpperCamelCase = { "A": "N", "N": "A", "B": "O", "O": "B", "C": "P", "P": "C", "D": "Q", "Q": "D", "E": "R", "R": "E", "F": "S", "S": "F", "G": "T", "T": "G", "H": "U", "U": "H", "I": "V", "V": "I", "J": "W", "W": "J", "K": "X", "X": "K", "L": "Y", "Y": "L", "M": "Z", "Z": "M", } # -------------------------- extra rotors -------------------------- UpperCamelCase = "RMDJXFUWGISLHVTCQNKYPBEZOA" UpperCamelCase = "SGLCPQWZHKXAREONTFBVIYJUDM" UpperCamelCase = "HVSICLTYKQUBXDWAJZOMFGPREN" UpperCamelCase = "RZWQHFMVDBKICJLNTUXAGYPSOE" UpperCamelCase = "LFKIJODBEGAMQPXVUHYSTCZRWN" UpperCamelCase = "KOAEGVDHXPQZMLFTYWJNBRCIUS" def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> tuple[RotorPositionT, RotorSelectionT, dict[str, str]]: # Checks if there are 3 unique rotors if (unique_rotsel := len(set(SCREAMING_SNAKE_CASE ) )) < 3: _lowercase : Optional[int] = F"""Please use 3 unique rotors (not {unique_rotsel})""" raise Exception(SCREAMING_SNAKE_CASE ) # Checks if rotor positions are valid _lowercase , _lowercase , _lowercase : int = rotpos if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : Dict = F"""First rotor position is not within range of 1..26 ({rotorposa}""" raise ValueError(SCREAMING_SNAKE_CASE ) if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : int = F"""Second rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(SCREAMING_SNAKE_CASE ) if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : str = F"""Third rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(SCREAMING_SNAKE_CASE ) # Validates string and returns dict _lowercase : Tuple = _plugboard(SCREAMING_SNAKE_CASE ) return rotpos, rotsel, pbdict def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> dict[str, str]: # tests the input string if it # a) is type string # b) has even length (so pairs can be made) if not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): _lowercase : Optional[int] = F"""Plugboard setting isn't type string ({type(SCREAMING_SNAKE_CASE )})""" raise TypeError(SCREAMING_SNAKE_CASE ) elif len(SCREAMING_SNAKE_CASE ) % 2 != 0: _lowercase : Optional[int] = F"""Odd number of symbols ({len(SCREAMING_SNAKE_CASE )})""" raise Exception(SCREAMING_SNAKE_CASE ) elif pbstring == "": return {} pbstring.replace(' ' , '' ) # Checks if all characters are unique _lowercase : Dict = set() for i in pbstring: if i not in abc: _lowercase : str = F"""'{i}' not in list of symbols""" raise Exception(SCREAMING_SNAKE_CASE ) elif i in tmppbl: _lowercase : int = F"""Duplicate symbol ({i})""" raise Exception(SCREAMING_SNAKE_CASE ) else: tmppbl.add(SCREAMING_SNAKE_CASE ) del tmppbl # Created the dictionary _lowercase : Optional[Any] = {} for j in range(0 , len(SCREAMING_SNAKE_CASE ) - 1 , 2 ): _lowercase : Dict = pbstring[j + 1] _lowercase : Union[str, Any] = pbstring[j] return pb def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = (rotora, rotora, rotora) , SCREAMING_SNAKE_CASE = "" , ) -> str: _lowercase : List[str] = text.upper() _lowercase , _lowercase , _lowercase : List[str] = _validator( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , plugb.upper() ) _lowercase , _lowercase , _lowercase : Optional[int] = rotor_position _lowercase , _lowercase , _lowercase : Union[str, Any] = rotor_selection rotorposa -= 1 rotorposa -= 1 rotorposa -= 1 _lowercase : Optional[int] = [] # encryption/decryption process -------------------------- for symbol in text: if symbol in abc: # 1st plugboard -------------------------- if symbol in plugboard: _lowercase : Dict = plugboard[symbol] # rotor ra -------------------------- _lowercase : Optional[Any] = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : Union[str, Any] = rotora[index % len(SCREAMING_SNAKE_CASE )] # rotor rb -------------------------- _lowercase : Tuple = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : str = rotora[index % len(SCREAMING_SNAKE_CASE )] # rotor rc -------------------------- _lowercase : List[Any] = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : List[str] = rotora[index % len(SCREAMING_SNAKE_CASE )] # reflector -------------------------- # this is the reason you don't need another machine to decipher _lowercase : List[str] = reflector[symbol] # 2nd rotors _lowercase : List[str] = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] _lowercase : Tuple = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] _lowercase : Dict = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] # 2nd plugboard if symbol in plugboard: _lowercase : int = plugboard[symbol] # moves/resets rotor positions rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : Any = 0 rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : int = 0 rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : Any = 0 # else: # pass # Error could be also raised # raise ValueError( # 'Invalid symbol('+repr(symbol)+')') result.append(SCREAMING_SNAKE_CASE ) return "".join(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": UpperCamelCase = "This is my Python script that emulates the Enigma machine from WWII." UpperCamelCase = (1, 1, 1) UpperCamelCase = "pictures" UpperCamelCase = (rotora, rotora, rotora) UpperCamelCase = enigma(message, rotor_pos, rotor_sel, pb) print("Encrypted message:", en) print("Decrypted message:", enigma(en, rotor_pos, rotor_sel, pb))
66
1
from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import * # noqa F403 else: from .scheduling_consistency_models import CMStochasticIterativeScheduler from .scheduling_ddim import DDIMScheduler from .scheduling_ddim_inverse import DDIMInverseScheduler from .scheduling_ddim_parallel import DDIMParallelScheduler from .scheduling_ddpm import DDPMScheduler from .scheduling_ddpm_parallel import DDPMParallelScheduler from .scheduling_deis_multistep import DEISMultistepScheduler from .scheduling_dpmsolver_multistep import DPMSolverMultistepScheduler from .scheduling_dpmsolver_multistep_inverse import DPMSolverMultistepInverseScheduler from .scheduling_dpmsolver_singlestep import DPMSolverSinglestepScheduler from .scheduling_euler_ancestral_discrete import EulerAncestralDiscreteScheduler from .scheduling_euler_discrete import EulerDiscreteScheduler from .scheduling_heun_discrete import HeunDiscreteScheduler from .scheduling_ipndm import IPNDMScheduler from .scheduling_k_dpm_2_ancestral_discrete import KDPMaAncestralDiscreteScheduler from .scheduling_k_dpm_2_discrete import KDPMaDiscreteScheduler from .scheduling_karras_ve import KarrasVeScheduler from .scheduling_pndm import PNDMScheduler from .scheduling_repaint import RePaintScheduler from .scheduling_sde_ve import ScoreSdeVeScheduler from .scheduling_sde_vp import ScoreSdeVpScheduler from .scheduling_unclip import UnCLIPScheduler from .scheduling_unipc_multistep import UniPCMultistepScheduler from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin from .scheduling_vq_diffusion import VQDiffusionScheduler try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_flax_objects import * # noqa F403 else: from .scheduling_ddim_flax import FlaxDDIMScheduler from .scheduling_ddpm_flax import FlaxDDPMScheduler from .scheduling_dpmsolver_multistep_flax import FlaxDPMSolverMultistepScheduler from .scheduling_karras_ve_flax import FlaxKarrasVeScheduler from .scheduling_lms_discrete_flax import FlaxLMSDiscreteScheduler from .scheduling_pndm_flax import FlaxPNDMScheduler from .scheduling_sde_ve_flax import FlaxScoreSdeVeScheduler from .scheduling_utils_flax import ( FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, broadcast_to_shape_from_left, ) try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .scheduling_lms_discrete import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .scheduling_dpmsolver_sde import DPMSolverSDEScheduler
66
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = {"configuration_glpn": ["GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP", "GLPNConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["GLPNFeatureExtractor"] UpperCamelCase = ["GLPNImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "GLPN_PRETRAINED_MODEL_ARCHIVE_LIST", "GLPNForDepthEstimation", "GLPNLayer", "GLPNModel", "GLPNPreTrainedModel", ] if TYPE_CHECKING: from .configuration_glpn import GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP, GLPNConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_glpn import GLPNFeatureExtractor from .image_processing_glpn import GLPNImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_glpn import ( GLPN_PRETRAINED_MODEL_ARCHIVE_LIST, GLPNForDepthEstimation, GLPNLayer, GLPNModel, GLPNPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
1
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): _lowercase : List[Any] = torch.nn.Linear(1_0 , 1_0 ) _lowercase : Any = torch.optim.SGD(model.parameters() , 0.1 ) _lowercase : str = Accelerator() _lowercase : Any = accelerator.prepare(_lowerCAmelCase ) try: pickle.loads(pickle.dumps(_lowerCAmelCase ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
66
import unittest from transformers import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device if is_torch_available(): import torch from transformers import AutoModelForImageClassification if is_vision_available(): from transformers import AutoImageProcessor @require_torch @require_vision class lowerCAmelCase_ ( unittest.TestCase ): @slow def __a ( self ): _lowercase : List[Any] = AutoImageProcessor.from_pretrained('microsoft/dit-base-finetuned-rvlcdip' ) _lowercase : List[Any] = AutoModelForImageClassification.from_pretrained('microsoft/dit-base-finetuned-rvlcdip' ) model.to(_lowerCAmelCase ) from datasets import load_dataset _lowercase : Union[str, Any] = load_dataset('nielsr/rvlcdip-demo' ) _lowercase : Any = dataset['train'][0]['image'].convert('RGB' ) _lowercase : List[str] = image_processor(_lowerCAmelCase , return_tensors='pt' ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): _lowercase : Dict = model(**_lowerCAmelCase ) _lowercase : Any = outputs.logits _lowercase : str = torch.Size((1, 1_6) ) self.assertEqual(logits.shape , _lowerCAmelCase ) _lowercase : Union[str, Any] = torch.tensor( [-0.41_58, -0.40_92, -0.43_47] , device=_lowerCAmelCase , dtype=torch.float , ) self.assertTrue(torch.allclose(logits[0, :3] , _lowerCAmelCase , atol=1E-4 ) )
66
1
import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : List[str] = BarthezTokenizer _UpperCamelCase : List[str] = BarthezTokenizerFast _UpperCamelCase : List[Any] = True _UpperCamelCase : Tuple = True def __a ( self ): super().setUp() _lowercase : int = BarthezTokenizerFast.from_pretrained('moussaKam/mbarthez' ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=_lowerCAmelCase ) _lowercase : Union[str, Any] = tokenizer def __a ( self ): _lowercase : Dict = '<pad>' _lowercase : Any = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_lowerCAmelCase ) , _lowerCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_lowerCAmelCase ) , _lowerCAmelCase ) def __a ( self ): _lowercase : Dict = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<s>' ) self.assertEqual(vocab_keys[1] , '<pad>' ) self.assertEqual(vocab_keys[-1] , '<mask>' ) self.assertEqual(len(_lowerCAmelCase ) , 1_0_1_1_2_2 ) def __a ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1_0_1_1_2_2 ) @require_torch def __a ( self ): _lowercase : Union[str, Any] = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] _lowercase : Optional[Any] = [0, 5_7, 3_0_1_8, 7_0_3_0_7, 9_1, 2] _lowercase : Any = self.tokenizer( _lowerCAmelCase , max_length=len(_lowerCAmelCase ) , padding=_lowerCAmelCase , truncation=_lowerCAmelCase , return_tensors='pt' ) self.assertIsInstance(_lowerCAmelCase , _lowerCAmelCase ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) _lowercase : Optional[int] = batch.input_ids.tolist()[0] self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) def __a ( self ): if not self.test_rust_tokenizer: return _lowercase : int = self.get_tokenizer() _lowercase : Tuple = self.get_rust_tokenizer() _lowercase : List[str] = 'I was born in 92000, and this is falsé.' _lowercase : str = tokenizer.tokenize(_lowerCAmelCase ) _lowercase : Optional[int] = rust_tokenizer.tokenize(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) _lowercase : Optional[Any] = tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) _lowercase : Tuple = rust_tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) _lowercase : Dict = self.get_rust_tokenizer() _lowercase : Optional[int] = tokenizer.encode(_lowerCAmelCase ) _lowercase : int = rust_tokenizer.encode(_lowerCAmelCase ) self.assertListEqual(_lowerCAmelCase , _lowerCAmelCase ) @slow def __a ( self ): # fmt: off _lowercase : List[str] = {'input_ids': [[0, 4_9_0, 1_4_3_2_8, 4_5_0_7, 3_5_4, 4_7, 4_3_6_6_9, 9_5, 2_5, 7_8_1_1_7, 2_0_2_1_5, 1_9_7_7_9, 1_9_0, 2_2, 4_0_0, 4, 3_5_3_4_3, 8_0_3_1_0, 6_0_3, 8_6, 2_4_9_3_7, 1_0_5, 3_3_4_3_8, 9_4_7_6_2, 1_9_6, 3_9_6_4_2, 7, 1_5, 1_5_9_3_3, 1_7_3, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_0_5_3_4, 8_7, 2_5, 6_6, 3_3_5_8, 1_9_6, 5_5_2_8_9, 8, 8_2_9_6_1, 8_1, 2_2_0_4, 7_5_2_0_3, 7, 1_5, 7_6_3, 1_2_9_5_6, 2_1_6, 1_7_8, 1_4_3_2_8, 9_5_9_5, 1_3_7_7, 6_9_6_9_3, 7, 4_4_8, 7_1_0_2_1, 1_9_6, 1_8_1_0_6, 1_4_3_7, 1_3_9_7_4, 1_0_8, 9_0_8_3, 4, 4_9_3_1_5, 7, 3_9, 8_6, 1_3_2_6, 2_7_9_3, 4_6_3_3_3, 4, 4_4_8, 1_9_6, 7_4_5_8_8, 7, 4_9_3_1_5, 7, 3_9, 2_1, 8_2_2, 3_8_4_7_0, 7_4, 2_1, 6_6_7_2_3, 6_2_4_8_0, 8, 2_2_0_5_0, 5, 2]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. _lowercase : Union[str, Any] = [ 'Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ' 'utilisé principalement dans le domaine du traitement automatique des langues (TAL).', 'À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ' 'pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ' 'telles que la traduction et la synthèse de texte.', ] self.tokenizer_integration_test_util( expected_encoding=_lowerCAmelCase , model_name='moussaKam/mbarthez' , revision='c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6' , sequences=_lowerCAmelCase , )
66
from PIL import Image def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Image: def brightness(SCREAMING_SNAKE_CASE ) -> float: return 128 + level + (c - 128) if not -255.0 <= level <= 255.0: raise ValueError('level must be between -255.0 (black) and 255.0 (white)' ) return img.point(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": # Load image with Image.open("image_data/lena.jpg") as img: # Change brightness to 100 UpperCamelCase = change_brightness(img, 100) brigt_img.save("image_data/lena_brightness.png", format="png")
66
1
def __magic_name__ ( SCREAMING_SNAKE_CASE = 1_000_000 ) -> int: _lowercase : List[str] = limit + 1 _lowercase : Dict = [0] * limit for first_term in range(1 , SCREAMING_SNAKE_CASE ): for n in range(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): _lowercase : Union[str, Any] = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a _lowercase : Union[str, Any] = sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(f'''{solution() = }''')
66
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): _lowercase : List[Any] = torch.nn.Linear(1_0 , 1_0 ) _lowercase : Any = torch.optim.SGD(model.parameters() , 0.1 ) _lowercase : str = Accelerator() _lowercase : Any = accelerator.prepare(_lowerCAmelCase ) try: pickle.loads(pickle.dumps(_lowerCAmelCase ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
66
1
import copy import os import cva import numpy as np from matplotlib import pyplot as plt class lowerCAmelCase_ : def __init__( self ): _lowercase : Optional[Any] = '' _lowercase : List[Any] = '' _lowercase : Union[str, Any] = [] _lowercase : List[str] = 0 _lowercase : Dict = 2_5_6 _lowercase : Optional[Any] = 0 _lowercase : List[str] = 0 _lowercase : Union[str, Any] = 0 _lowercase : str = 0 def __a ( self , _lowerCAmelCase ): _lowercase : Union[str, Any] = cva.imread(_lowerCAmelCase , 0 ) _lowercase : Tuple = copy.deepcopy(self.img ) _lowercase , _lowercase , _lowercase : Tuple = plt.hist(self.img.ravel() , 2_5_6 , [0, 2_5_6] , label='x' ) _lowercase : int = np.sum(_lowerCAmelCase ) for i in range(len(_lowerCAmelCase ) ): _lowercase : Dict = x[i] / self.k self.sk += prk _lowercase : int = (self.L - 1) * self.sk if self.rem != 0: _lowercase : Union[str, Any] = int(last % last ) _lowercase : int = int(last + 1 if self.rem >= 0.5 else last ) self.last_list.append(_lowerCAmelCase ) _lowercase : List[Any] = int(np.ma.count(self.img ) / self.img[1].size ) _lowercase : Dict = self.img[1].size for i in range(self.number_of_cols ): for j in range(self.number_of_rows ): _lowercase : Any = self.img[j][i] if num != self.last_list[num]: _lowercase : Union[str, Any] = self.last_list[num] cva.imwrite('output_data/output.jpg' , self.img ) def __a ( self ): plt.hist(self.img.ravel() , 2_5_6 , [0, 2_5_6] ) def __a ( self ): cva.imshow('Output-Image' , self.img ) cva.imshow('Input-Image' , self.original_image ) cva.waitKey(5_0_0_0 ) cva.destroyAllWindows() if __name__ == "__main__": UpperCamelCase = os.path.join(os.path.basename(__file__), "image_data/input.jpg") UpperCamelCase = ConstantStretch() stretcher.stretch(file_path) stretcher.plot_histogram() stretcher.show_image()
66
import requests from bsa import BeautifulSoup def __magic_name__ ( SCREAMING_SNAKE_CASE = "AAPL" ) -> str: _lowercase : str = F"""https://in.finance.yahoo.com/quote/{symbol}?s={symbol}""" _lowercase : int = BeautifulSoup(requests.get(SCREAMING_SNAKE_CASE ).text , 'html.parser' ) _lowercase : List[str] = 'My(6px) Pos(r) smartphone_Mt(6px)' return soup.find('div' , class_=class_ ).find('span' ).text if __name__ == "__main__": for symbol in "AAPL AMZN IBM GOOG MSFT ORCL".split(): print(f'''Current {symbol:<4} stock price is {stock_price(symbol):>8}''')
66
1
import collections import inspect import unittest from transformers import FocalNetConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( FocalNetBackbone, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetModel, ) from transformers.models.focalnet.modeling_focalnet import FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowerCAmelCase_ : def __init__( self , _lowerCAmelCase , _lowerCAmelCase=1_3 , _lowerCAmelCase=3_2 , _lowerCAmelCase=2 , _lowerCAmelCase=3 , _lowerCAmelCase=1_6 , _lowerCAmelCase=[3_2, 6_4, 1_2_8] , _lowerCAmelCase=[1, 2, 1] , _lowerCAmelCase=[2, 2, 4] , _lowerCAmelCase=2 , _lowerCAmelCase=2.0 , _lowerCAmelCase=True , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.1 , _lowerCAmelCase="gelu" , _lowerCAmelCase=False , _lowerCAmelCase=True , _lowerCAmelCase=0.02 , _lowerCAmelCase=1E-5 , _lowerCAmelCase=True , _lowerCAmelCase=None , _lowerCAmelCase=True , _lowerCAmelCase=1_0 , _lowerCAmelCase=8 , _lowerCAmelCase=["stage1", "stage2"] , _lowerCAmelCase=[1, 2] , ): _lowercase : int = parent _lowercase : str = batch_size _lowercase : Optional[Any] = image_size _lowercase : Union[str, Any] = patch_size _lowercase : Optional[Any] = num_channels _lowercase : List[str] = embed_dim _lowercase : Optional[Any] = hidden_sizes _lowercase : Dict = depths _lowercase : int = num_heads _lowercase : Union[str, Any] = window_size _lowercase : Optional[int] = mlp_ratio _lowercase : Dict = qkv_bias _lowercase : Tuple = hidden_dropout_prob _lowercase : Optional[int] = attention_probs_dropout_prob _lowercase : Any = drop_path_rate _lowercase : Optional[Any] = hidden_act _lowercase : List[Any] = use_absolute_embeddings _lowercase : List[Any] = patch_norm _lowercase : Optional[Any] = layer_norm_eps _lowercase : Optional[Any] = initializer_range _lowercase : int = is_training _lowercase : List[str] = scope _lowercase : List[str] = use_labels _lowercase : Tuple = type_sequence_label_size _lowercase : str = encoder_stride _lowercase : Dict = out_features _lowercase : Optional[int] = out_indices def __a ( self ): _lowercase : Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowercase : Dict = None if self.use_labels: _lowercase : Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowercase : List[Any] = self.get_config() return config, pixel_values, labels def __a ( self ): return FocalNetConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , hidden_sizes=self.hidden_sizes , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): _lowercase : Optional[Any] = FocalNetModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() _lowercase : int = model(_lowerCAmelCase ) _lowercase : List[str] = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) _lowercase : Any = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): _lowercase : Optional[int] = FocalNetBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() _lowercase : Dict = model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.image_size, 8, 8] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[:-1] ) # verify backbone works with out_features=None _lowercase : List[str] = None _lowercase : Optional[Any] = FocalNetBackbone(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() _lowercase : int = model(_lowerCAmelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.image_size * 2, 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): _lowercase : Any = FocalNetForMaskedImageModeling(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() _lowercase : Union[str, Any] = model(_lowerCAmelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images _lowercase : Optional[int] = 1 _lowercase : List[str] = FocalNetForMaskedImageModeling(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() _lowercase : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _lowercase : Optional[int] = model(_lowerCAmelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): _lowercase : Optional[int] = self.type_sequence_label_size _lowercase : str = FocalNetForImageClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() _lowercase : Any = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images _lowercase : Dict = 1 _lowercase : Tuple = FocalNetForImageClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() _lowercase : List[Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _lowercase : int = model(_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __a ( self ): _lowercase : int = self.prepare_config_and_inputs() _lowercase , _lowercase , _lowercase : Tuple = config_and_inputs _lowercase : Dict = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class lowerCAmelCase_ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCamelCase : int = ( ( FocalNetModel, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetBackbone, ) if is_torch_available() else () ) _UpperCamelCase : List[str] = ( {"feature-extraction": FocalNetModel, "image-classification": FocalNetForImageClassification} if is_torch_available() else {} ) _UpperCamelCase : int = False _UpperCamelCase : Any = False _UpperCamelCase : List[str] = False _UpperCamelCase : Union[str, Any] = False _UpperCamelCase : Optional[int] = False def __a ( self ): _lowercase : int = FocalNetModelTester(self ) _lowercase : int = ConfigTester(self , config_class=_lowerCAmelCase , embed_dim=3_7 , has_text_modality=_lowerCAmelCase ) def __a ( self ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __a ( self ): return def __a ( self ): _lowercase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def __a ( self ): _lowercase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_lowerCAmelCase ) def __a ( self ): _lowercase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_lowerCAmelCase ) def __a ( self ): _lowercase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) @unittest.skip(reason='FocalNet does not use inputs_embeds' ) def __a ( self ): pass @unittest.skip(reason='FocalNet does not use feedforward chunking' ) def __a ( self ): pass def __a ( self ): _lowercase , _lowercase : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes[:-1]: _lowercase : Optional[Any] = model_class(_lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _lowercase : Any = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCAmelCase , nn.Linear ) ) def __a ( self ): _lowercase , _lowercase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes[:-1]: _lowercase : Union[str, Any] = model_class(_lowerCAmelCase ) _lowercase : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowercase : str = [*signature.parameters.keys()] _lowercase : int = ['pixel_values'] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): _lowercase : List[Any] = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): _lowercase : Any = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) _lowercase : Any = outputs.hidden_states _lowercase : List[Any] = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(_lowerCAmelCase ) , _lowerCAmelCase ) # FocalNet has a different seq_length _lowercase : Tuple = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) _lowercase : Union[str, Any] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) _lowercase : Union[str, Any] = outputs.reshaped_hidden_states self.assertEqual(len(_lowerCAmelCase ) , _lowerCAmelCase ) _lowercase , _lowercase , _lowercase , _lowercase : Optional[int] = reshaped_hidden_states[0].shape _lowercase : Optional[int] = ( reshaped_hidden_states[0].view(_lowerCAmelCase , _lowerCAmelCase , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def __a ( self ): _lowercase , _lowercase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() _lowercase : Optional[int] = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes[:-1]: _lowercase : Tuple = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowercase : Dict = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def __a ( self ): _lowercase , _lowercase : Any = self.model_tester.prepare_config_and_inputs_for_common() _lowercase : Optional[Any] = 3 _lowercase : Tuple = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) _lowercase : Tuple = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) _lowercase : Any = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) _lowercase : Union[str, Any] = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes[:-1]: _lowercase : List[Any] = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowercase : Optional[Any] = True self.check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , (padded_height, padded_width) ) @slow def __a ( self ): for model_name in FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowercase : List[str] = FocalNetModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def __a ( self ): _lowercase , _lowercase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() _lowercase : Tuple = _config_zero_init(_lowerCAmelCase ) for model_class in self.all_model_classes: _lowercase : Optional[Any] = model_class(config=_lowerCAmelCase ) for name, param in model.named_parameters(): if "embeddings" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @require_vision @require_torch class lowerCAmelCase_ ( unittest.TestCase ): @cached_property def __a ( self ): # TODO update organization return AutoImageProcessor.from_pretrained('microsoft/focalnet-tiny' ) if is_vision_available() else None @slow def __a ( self ): _lowercase : str = FocalNetForImageClassification.from_pretrained('microsoft/focalnet-tiny' ).to(_lowerCAmelCase ) _lowercase : Optional[int] = self.default_image_processor _lowercase : Tuple = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) _lowercase : List[Any] = image_processor(images=_lowerCAmelCase , return_tensors='pt' ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): _lowercase : int = model(**_lowerCAmelCase ) # verify the logits _lowercase : Dict = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) _lowercase : Any = torch.tensor([0.21_66, -0.43_68, 0.21_91] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCAmelCase , atol=1E-4 ) ) self.assertTrue(outputs.logits.argmax(dim=-1 ).item() , 2_8_1 ) @require_torch class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : List[str] = (FocalNetBackbone,) if is_torch_available() else () _UpperCamelCase : Dict = FocalNetConfig _UpperCamelCase : Optional[Any] = False def __a ( self ): _lowercase : str = FocalNetModelTester(self )
66
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = { "configuration_poolformer": [ "POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "PoolFormerConfig", "PoolFormerOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["PoolFormerFeatureExtractor"] UpperCamelCase = ["PoolFormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "PoolFormerForImageClassification", "PoolFormerModel", "PoolFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure)
66
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCamelCase = { "configuration_jukebox": [ "JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP", "JukeboxConfig", "JukeboxPriorConfig", "JukeboxVQVAEConfig", ], "tokenization_jukebox": ["JukeboxTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST", "JukeboxModel", "JukeboxPreTrainedModel", "JukeboxVQVAE", "JukeboxPrior", ] if TYPE_CHECKING: from .configuration_jukebox import ( JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP, JukeboxConfig, JukeboxPriorConfig, JukeboxVQVAEConfig, ) from .tokenization_jukebox import JukeboxTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_jukebox import ( JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST, JukeboxModel, JukeboxPreTrainedModel, JukeboxPrior, JukeboxVQVAE, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : Dict = LayoutLMTokenizer _UpperCamelCase : Union[str, Any] = LayoutLMTokenizerFast _UpperCamelCase : int = True _UpperCamelCase : Tuple = True def __a ( self ): super().setUp() _lowercase : Union[str, Any] = [ '[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] _lowercase : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def __a ( self , **_lowerCAmelCase ): return LayoutLMTokenizer.from_pretrained(self.tmpdirname , **_lowerCAmelCase ) def __a ( self , _lowerCAmelCase ): _lowercase : str = 'UNwant\u00E9d,running' _lowercase : List[Any] = 'unwanted, running' return input_text, output_text def __a ( self ): _lowercase : Dict = self.tokenizer_class(self.vocab_file ) _lowercase : Dict = tokenizer.tokenize('UNwant\u00E9d,running' ) self.assertListEqual(_lowerCAmelCase , ['un', '##want', '##ed', ',', 'runn', '##ing'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [7, 4, 5, 1_0, 8, 9] ) def __a ( self ): pass
66
1
from __future__ import annotations from typing import Any class lowerCAmelCase_ ( __snake_case ): pass class lowerCAmelCase_ : def __init__( self , _lowerCAmelCase ): _lowercase : Any = data _lowercase : Node | None = None def __iter__( self ): _lowercase : Union[str, Any] = self _lowercase : Tuple = [] while node: if node in visited: raise ContainsLoopError visited.append(_lowerCAmelCase ) yield node.data _lowercase : Optional[int] = node.next_node @property def __a ( self ): try: list(self ) return False except ContainsLoopError: return True if __name__ == "__main__": UpperCamelCase = Node(1) UpperCamelCase = Node(2) UpperCamelCase = Node(3) UpperCamelCase = Node(4) print(root_node.has_loop) # False UpperCamelCase = root_node.next_node print(root_node.has_loop) # True UpperCamelCase = Node(5) UpperCamelCase = Node(6) UpperCamelCase = Node(5) UpperCamelCase = Node(6) print(root_node.has_loop) # False UpperCamelCase = Node(1) print(root_node.has_loop) # False
66
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : str = ShapEPipeline _UpperCamelCase : Any = ["prompt"] _UpperCamelCase : int = ["prompt"] _UpperCamelCase : Union[str, Any] = [ "num_images_per_prompt", "num_inference_steps", "generator", "latents", "guidance_scale", "frame_size", "output_type", "return_dict", ] _UpperCamelCase : Optional[Any] = False @property def __a ( self ): return 3_2 @property def __a ( self ): return 3_2 @property def __a ( self ): return self.time_input_dim * 4 @property def __a ( self ): return 8 @property def __a ( self ): _lowercase : Any = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) return tokenizer @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) return CLIPTextModelWithProjection(_lowerCAmelCase ) @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = { 'num_attention_heads': 2, 'attention_head_dim': 1_6, 'embedding_dim': self.time_input_dim, 'num_embeddings': 3_2, 'embedding_proj_dim': self.text_embedder_hidden_size, 'time_embed_dim': self.time_embed_dim, 'num_layers': 1, 'clip_embed_dim': self.time_input_dim * 2, 'additional_embeddings': 0, 'time_embed_act_fn': 'gelu', 'norm_in_type': 'layer', 'encoder_hid_proj_type': None, 'added_emb_type': None, } _lowercase : Optional[Any] = PriorTransformer(**_lowerCAmelCase ) return model @property def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = { 'param_shapes': ( (self.renderer_dim, 9_3), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), 'd_latent': self.time_input_dim, 'd_hidden': self.renderer_dim, 'n_output': 1_2, 'background': ( 0.1, 0.1, 0.1, ), } _lowercase : List[Any] = ShapERenderer(**_lowerCAmelCase ) return model def __a ( self ): _lowercase : Optional[Any] = self.dummy_prior _lowercase : Dict = self.dummy_text_encoder _lowercase : List[str] = self.dummy_tokenizer _lowercase : Union[str, Any] = self.dummy_renderer _lowercase : List[str] = HeunDiscreteScheduler( beta_schedule='exp' , num_train_timesteps=1_0_2_4 , prediction_type='sample' , use_karras_sigmas=_lowerCAmelCase , clip_sample=_lowerCAmelCase , clip_sample_range=1.0 , ) _lowercase : List[str] = { 'prior': prior, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'renderer': renderer, 'scheduler': scheduler, } return components def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : Optional[Any] = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Optional[int] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : List[Any] = { 'prompt': 'horse', 'generator': generator, 'num_inference_steps': 1, 'frame_size': 3_2, 'output_type': 'np', } return inputs def __a ( self ): _lowercase : Optional[int] = 'cpu' _lowercase : List[Any] = self.get_dummy_components() _lowercase : Tuple = self.pipeline_class(**_lowerCAmelCase ) _lowercase : Union[str, Any] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Union[str, Any] = pipe(**self.get_dummy_inputs(_lowerCAmelCase ) ) _lowercase : str = output.images[0] _lowercase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (2_0, 3_2, 3_2, 3) _lowercase : str = np.array( [ 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __a ( self ): # NOTE: Larger batch sizes cause this test to timeout, only test on smaller batches self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def __a ( self ): _lowercase : List[Any] = torch_device == 'cpu' _lowercase : Any = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=_lowerCAmelCase , relax_max_difference=_lowerCAmelCase , ) def __a ( self ): _lowercase : Union[str, Any] = self.get_dummy_components() _lowercase : Optional[int] = self.pipeline_class(**_lowerCAmelCase ) _lowercase : Any = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : str = 1 _lowercase : Optional[int] = 2 _lowercase : List[str] = self.get_dummy_inputs(_lowerCAmelCase ) for key in inputs.keys(): if key in self.batch_params: _lowercase : int = batch_size * [inputs[key]] _lowercase : Optional[int] = pipe(**_lowerCAmelCase , num_images_per_prompt=_lowerCAmelCase )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ): _lowercase : List[str] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/shap_e/test_shap_e_np_out.npy' ) _lowercase : Any = ShapEPipeline.from_pretrained('openai/shap-e' ) _lowercase : List[str] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Tuple = torch.Generator(device=_lowerCAmelCase ).manual_seed(0 ) _lowercase : int = pipe( 'a shark' , generator=_lowerCAmelCase , guidance_scale=15.0 , num_inference_steps=6_4 , frame_size=6_4 , output_type='np' , ).images[0] assert images.shape == (2_0, 6_4, 6_4, 3) assert_mean_pixel_difference(_lowerCAmelCase , _lowerCAmelCase )
66
1
from __future__ import annotations def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None ) -> list[list[str]]: _lowercase : Union[str, Any] = word_bank or [] # create a table _lowercase : int = len(SCREAMING_SNAKE_CASE ) + 1 _lowercase : list[list[list[str]]] = [] for _ in range(SCREAMING_SNAKE_CASE ): table.append([] ) # seed value _lowercase : int = [[]] # because empty string has empty combination # iterate through the indices for i in range(SCREAMING_SNAKE_CASE ): # condition if table[i] != []: for word in word_bank: # slice condition if target[i : i + len(SCREAMING_SNAKE_CASE )] == word: _lowercase : list[list[str]] = [ [word, *way] for way in table[i] ] # adds the word to every combination the current position holds # now,push that combination to the table[i+len(word)] table[i + len(SCREAMING_SNAKE_CASE )] += new_combinations # combinations are in reverse order so reverse for better output for combination in table[len(SCREAMING_SNAKE_CASE )]: combination.reverse() return table[len(SCREAMING_SNAKE_CASE )] if __name__ == "__main__": print(all_construct("jwajalapa", ["jwa", "j", "w", "a", "la", "lapa"])) print(all_construct("rajamati", ["s", "raj", "amat", "raja", "ma", "i", "t"])) print( all_construct( "hexagonosaurus", ["h", "ex", "hex", "ag", "ago", "ru", "auru", "rus", "go", "no", "o", "s"], ) )
66
import sys UpperCamelCase = ( "73167176531330624919225119674426574742355349194934" "96983520312774506326239578318016984801869478851843" "85861560789112949495459501737958331952853208805511" "12540698747158523863050715693290963295227443043557" "66896648950445244523161731856403098711121722383113" "62229893423380308135336276614282806444486645238749" "30358907296290491560440772390713810515859307960866" "70172427121883998797908792274921901699720888093776" "65727333001053367881220235421809751254540594752243" "52584907711670556013604839586446706324415722155397" "53697817977846174064955149290862569321978468622482" "83972241375657056057490261407972968652414535100474" "82166370484403199890008895243450658541227588666881" "16427171479924442928230863465674813919123162824586" "17866458359124566529476545682848912883142607690042" "24219022671055626321111109370544217506941658960408" "07198403850962455444362981230987879927244284909188" "84580156166097919133875499200524063689912560717606" "05886116467109405077541002256983155200055935729725" "71636269561882670428252483600823257530420752963450" ) def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : List[Any] = 1 for digit in s: product *= int(SCREAMING_SNAKE_CASE ) return product def __magic_name__ ( SCREAMING_SNAKE_CASE = N ) -> int: _lowercase : Dict = -sys.maxsize - 1 _lowercase : Tuple = n[:13] _lowercase : List[Any] = 13 while cur_index < len(SCREAMING_SNAKE_CASE ) - 13: if int(n[cur_index] ) >= int(substr[0] ): _lowercase : List[str] = substr[1:] + n[cur_index] cur_index += 1 else: _lowercase : str = max(SCREAMING_SNAKE_CASE , str_eval(SCREAMING_SNAKE_CASE ) ) _lowercase : Dict = n[cur_index : cur_index + 13] cur_index += 13 return largest_product if __name__ == "__main__": print(f'''{solution() = }''')
66
1
from scipy.stats import pearsonr, spearmanr from sklearn.metrics import fa_score, matthews_corrcoef import datasets UpperCamelCase = "\\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n" UpperCamelCase = "\\nGLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n" UpperCamelCase = "\nCompute GLUE evaluation metric associated to each GLUE dataset.\nArgs:\n predictions: list of predictions to score.\n Each translation should be tokenized into a list of tokens.\n references: list of lists of references for each translation.\n Each reference should be tokenized into a list of tokens.\nReturns: depending on the GLUE subset, one or several of:\n \"accuracy\": Accuracy\n \"f1\": F1 score\n \"pearson\": Pearson Correlation\n \"spearmanr\": Spearman Correlation\n \"matthews_correlation\": Matthew Correlation\nExamples:\n\n >>> glue_metric = datasets.load_metric('glue', 'sst2') # 'sst2' or any of [\"mnli\", \"mnli_mismatched\", \"mnli_matched\", \"qnli\", \"rte\", \"wnli\", \"hans\"]\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'accuracy': 1.0}\n\n >>> glue_metric = datasets.load_metric('glue', 'mrpc') # 'mrpc' or 'qqp'\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'accuracy': 1.0, 'f1': 1.0}\n\n >>> glue_metric = datasets.load_metric('glue', 'stsb')\n >>> references = [0., 1., 2., 3., 4., 5.]\n >>> predictions = [0., 1., 2., 3., 4., 5.]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print({\"pearson\": round(results[\"pearson\"], 2), \"spearmanr\": round(results[\"spearmanr\"], 2)})\n {'pearson': 1.0, 'spearmanr': 1.0}\n\n >>> glue_metric = datasets.load_metric('glue', 'cola')\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'matthews_correlation': 1.0}\n" def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> int: return float((preds == labels).mean() ) def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Optional[Any]: _lowercase : Optional[int] = simple_accuracy(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) _lowercase : int = float(fa_score(y_true=SCREAMING_SNAKE_CASE , y_pred=SCREAMING_SNAKE_CASE ) ) return { "accuracy": acc, "f1": fa, } def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Any: _lowercase : Optional[Any] = float(pearsonr(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE )[0] ) _lowercase : int = float(spearmanr(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE )[0] ) return { "pearson": pearson_corr, "spearmanr": spearman_corr, } @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase_ ( datasets.Metric ): def __a ( self ): if self.config_name not in [ "sst2", "mnli", "mnli_mismatched", "mnli_matched", "cola", "stsb", "mrpc", "qqp", "qnli", "rte", "wnli", "hans", ]: raise KeyError( 'You should supply a configuration name selected in ' '["sst2", "mnli", "mnli_mismatched", "mnli_matched", ' '"cola", "stsb", "mrpc", "qqp", "qnli", "rte", "wnli", "hans"]' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('int64' if self.config_name != 'stsb' else 'float32' ), 'references': datasets.Value('int64' if self.config_name != 'stsb' else 'float32' ), } ) , codebase_urls=[] , reference_urls=[] , format='numpy' , ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase ): if self.config_name == "cola": return {"matthews_correlation": matthews_corrcoef(_lowerCAmelCase , _lowerCAmelCase )} elif self.config_name == "stsb": return pearson_and_spearman(_lowerCAmelCase , _lowerCAmelCase ) elif self.config_name in ["mrpc", "qqp"]: return acc_and_fa(_lowerCAmelCase , _lowerCAmelCase ) elif self.config_name in ["sst2", "mnli", "mnli_mismatched", "mnli_matched", "qnli", "rte", "wnli", "hans"]: return {"accuracy": simple_accuracy(_lowerCAmelCase , _lowerCAmelCase )} else: raise KeyError( 'You should supply a configuration name selected in ' '["sst2", "mnli", "mnli_mismatched", "mnli_matched", ' '"cola", "stsb", "mrpc", "qqp", "qnli", "rte", "wnli", "hans"]' )
66
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available UpperCamelCase = { "configuration_gpt_neo": ["GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTNeoConfig", "GPTNeoOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTNeoForCausalLM", "GPTNeoForQuestionAnswering", "GPTNeoForSequenceClassification", "GPTNeoForTokenClassification", "GPTNeoModel", "GPTNeoPreTrainedModel", "load_tf_weights_in_gpt_neo", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "FlaxGPTNeoForCausalLM", "FlaxGPTNeoModel", "FlaxGPTNeoPreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_neo import GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoConfig, GPTNeoOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neo import ( GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoForCausalLM, GPTNeoForQuestionAnswering, GPTNeoForSequenceClassification, GPTNeoForTokenClassification, GPTNeoModel, GPTNeoPreTrainedModel, load_tf_weights_in_gpt_neo, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_gpt_neo import FlaxGPTNeoForCausalLM, FlaxGPTNeoModel, FlaxGPTNeoPreTrainedModel else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = { "microsoft/swin-tiny-patch4-window7-224": ( "https://huggingface.co/microsoft/swin-tiny-patch4-window7-224/resolve/main/config.json" ), # See all Swin models at https://huggingface.co/models?filter=swin } class lowerCAmelCase_ ( __snake_case , __snake_case ): _UpperCamelCase : List[Any] = "swin" _UpperCamelCase : Tuple = { "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self , _lowerCAmelCase=2_2_4 , _lowerCAmelCase=4 , _lowerCAmelCase=3 , _lowerCAmelCase=9_6 , _lowerCAmelCase=[2, 2, 6, 2] , _lowerCAmelCase=[3, 6, 1_2, 2_4] , _lowerCAmelCase=7 , _lowerCAmelCase=4.0 , _lowerCAmelCase=True , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.1 , _lowerCAmelCase="gelu" , _lowerCAmelCase=False , _lowerCAmelCase=0.02 , _lowerCAmelCase=1E-5 , _lowerCAmelCase=3_2 , _lowerCAmelCase=None , _lowerCAmelCase=None , **_lowerCAmelCase , ): super().__init__(**_lowerCAmelCase ) _lowercase : Dict = image_size _lowercase : Optional[int] = patch_size _lowercase : str = num_channels _lowercase : List[Any] = embed_dim _lowercase : Optional[int] = depths _lowercase : Any = len(_lowerCAmelCase ) _lowercase : Tuple = num_heads _lowercase : List[Any] = window_size _lowercase : int = mlp_ratio _lowercase : Any = qkv_bias _lowercase : Any = hidden_dropout_prob _lowercase : Optional[Any] = attention_probs_dropout_prob _lowercase : List[str] = drop_path_rate _lowercase : Dict = hidden_act _lowercase : Tuple = use_absolute_embeddings _lowercase : Any = layer_norm_eps _lowercase : Dict = initializer_range _lowercase : Dict = encoder_stride # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model _lowercase : int = int(embed_dim * 2 ** (len(_lowerCAmelCase ) - 1) ) _lowercase : str = ['stem'] + [F"""stage{idx}""" for idx in range(1 , len(_lowerCAmelCase ) + 1 )] _lowercase , _lowercase : Any = get_aligned_output_features_output_indices( out_features=_lowerCAmelCase , out_indices=_lowerCAmelCase , stage_names=self.stage_names ) class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : str = version.parse("1.11" ) @property def __a ( self ): return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def __a ( self ): return 1E-4
66
from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : str = ["image_processor", "tokenizer"] _UpperCamelCase : Union[str, Any] = "AutoImageProcessor" _UpperCamelCase : Union[str, Any] = "AutoTokenizer" def __init__( self , _lowerCAmelCase , _lowerCAmelCase ): super().__init__(_lowerCAmelCase , _lowerCAmelCase ) _lowercase : Union[str, Any] = self.image_processor def __call__( self , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=None , **_lowerCAmelCase ): if text is None and images is None: raise ValueError('You have to specify either text or images. Both cannot be none.' ) if text is not None: _lowercase : Dict = self.tokenizer(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if images is not None: _lowercase : Union[str, Any] = self.image_processor(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if text is not None and images is not None: _lowercase : Optional[Any] = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_lowerCAmelCase ) , tensor_type=_lowerCAmelCase ) def __a ( self , *_lowerCAmelCase , **_lowerCAmelCase ): return self.tokenizer.batch_decode(*_lowerCAmelCase , **_lowerCAmelCase ) def __a ( self , *_lowerCAmelCase , **_lowerCAmelCase ): return self.tokenizer.decode(*_lowerCAmelCase , **_lowerCAmelCase ) @property def __a ( self ): return ["input_ids", "attention_mask", "pixel_values"]
66
1
from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : str = ["image_processor", "tokenizer"] _UpperCamelCase : Union[str, Any] = "AutoImageProcessor" _UpperCamelCase : Union[str, Any] = "AutoTokenizer" def __init__( self , _lowerCAmelCase , _lowerCAmelCase ): super().__init__(_lowerCAmelCase , _lowerCAmelCase ) _lowercase : Union[str, Any] = self.image_processor def __call__( self , _lowerCAmelCase=None , _lowerCAmelCase=None , _lowerCAmelCase=None , **_lowerCAmelCase ): if text is None and images is None: raise ValueError('You have to specify either text or images. Both cannot be none.' ) if text is not None: _lowercase : Dict = self.tokenizer(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if images is not None: _lowercase : Union[str, Any] = self.image_processor(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if text is not None and images is not None: _lowercase : Optional[Any] = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_lowerCAmelCase ) , tensor_type=_lowerCAmelCase ) def __a ( self , *_lowerCAmelCase , **_lowerCAmelCase ): return self.tokenizer.batch_decode(*_lowerCAmelCase , **_lowerCAmelCase ) def __a ( self , *_lowerCAmelCase , **_lowerCAmelCase ): return self.tokenizer.decode(*_lowerCAmelCase , **_lowerCAmelCase ) @property def __a ( self ): return ["input_ids", "attention_mask", "pixel_values"]
66
from __future__ import annotations import math def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> list[int]: if num <= 0: _lowercase : List[str] = F"""{num}: Invalid input, please enter a positive integer.""" raise ValueError(SCREAMING_SNAKE_CASE ) _lowercase : Union[str, Any] = [True] * (num + 1) _lowercase : Union[str, Any] = [] _lowercase : Dict = 2 _lowercase : Union[str, Any] = int(math.sqrt(SCREAMING_SNAKE_CASE ) ) while start <= end: # If start is a prime if sieve[start] is True: prime.append(SCREAMING_SNAKE_CASE ) # Set multiples of start be False for i in range(start * start , num + 1 , SCREAMING_SNAKE_CASE ): if sieve[i] is True: _lowercase : str = False start += 1 for j in range(end + 1 , num + 1 ): if sieve[j] is True: prime.append(SCREAMING_SNAKE_CASE ) return prime if __name__ == "__main__": print(prime_sieve(int(input("Enter a positive integer: ").strip())))
66
1
import numpy as np UpperCamelCase = [ ["a", "b", "c", "d", "e"], ["f", "g", "h", "i", "k"], ["l", "m", "n", "o", "p"], ["q", "r", "s", "t", "u"], ["v", "w", "x", "y", "z"], ] class lowerCAmelCase_ : def __init__( self ): _lowercase : Any = np.array(_lowerCAmelCase ) def __a ( self , _lowerCAmelCase ): _lowercase , _lowercase : List[str] = np.where(letter == self.SQUARE ) _lowercase : Optional[int] = np.concatenate([indexa + 1, indexa + 1] ) return indexes def __a ( self , _lowerCAmelCase , _lowerCAmelCase ): _lowercase : int = self.SQUARE[indexa - 1, indexa - 1] return letter def __a ( self , _lowerCAmelCase ): _lowercase : int = message.lower() _lowercase : Optional[Any] = message.replace(' ' , '' ) _lowercase : Optional[Any] = message.replace('j' , 'i' ) _lowercase : Optional[int] = np.empty((2, len(_lowerCAmelCase )) ) for letter_index in range(len(_lowerCAmelCase ) ): _lowercase : Any = self.letter_to_numbers(message[letter_index] ) _lowercase : Tuple = numbers[0] _lowercase : int = numbers[1] _lowercase : int = first_step.reshape(2 * len(_lowerCAmelCase ) ) _lowercase : Tuple = '' for numbers_index in range(len(_lowerCAmelCase ) ): _lowercase : int = int(second_step[numbers_index * 2] ) _lowercase : List[Any] = int(second_step[(numbers_index * 2) + 1] ) _lowercase : Optional[int] = self.numbers_to_letter(_lowerCAmelCase , _lowerCAmelCase ) _lowercase : List[Any] = encoded_message + letter return encoded_message def __a ( self , _lowerCAmelCase ): _lowercase : Tuple = message.lower() message.replace(' ' , '' ) _lowercase : Optional[Any] = np.empty(2 * len(_lowerCAmelCase ) ) for letter_index in range(len(_lowerCAmelCase ) ): _lowercase : List[Any] = self.letter_to_numbers(message[letter_index] ) _lowercase : int = numbers[0] _lowercase : Union[str, Any] = numbers[1] _lowercase : Optional[int] = first_step.reshape((2, len(_lowerCAmelCase )) ) _lowercase : Tuple = '' for numbers_index in range(len(_lowerCAmelCase ) ): _lowercase : Tuple = int(second_step[0, numbers_index] ) _lowercase : Dict = int(second_step[1, numbers_index] ) _lowercase : List[Any] = self.numbers_to_letter(_lowerCAmelCase , _lowerCAmelCase ) _lowercase : Dict = decoded_message + letter return decoded_message
66
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ConvNextConfig, SegformerImageProcessor, UperNetConfig, UperNetForSemanticSegmentation def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> List[Any]: _lowercase : int = 384 if "tiny" in model_name: _lowercase : Tuple = [3, 3, 9, 3] _lowercase : List[str] = [96, 192, 384, 768] if "small" in model_name: _lowercase : List[str] = [3, 3, 27, 3] _lowercase : Union[str, Any] = [96, 192, 384, 768] if "base" in model_name: _lowercase : List[Any] = [3, 3, 27, 3] _lowercase : Dict = [128, 256, 512, 1_024] _lowercase : Optional[int] = 512 if "large" in model_name: _lowercase : List[str] = [3, 3, 27, 3] _lowercase : List[Any] = [192, 384, 768, 1_536] _lowercase : Tuple = 768 if "xlarge" in model_name: _lowercase : str = [3, 3, 27, 3] _lowercase : List[str] = [256, 512, 1_024, 2_048] _lowercase : Tuple = 1_024 # set label information _lowercase : Dict = 150 _lowercase : Union[str, Any] = 'huggingface/label-files' _lowercase : str = 'ade20k-id2label.json' _lowercase : List[Any] = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , repo_type='dataset' ) , 'r' ) ) _lowercase : Dict = {int(SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} _lowercase : Tuple = {v: k for k, v in idalabel.items()} _lowercase : List[str] = ConvNextConfig( depths=SCREAMING_SNAKE_CASE , hidden_sizes=SCREAMING_SNAKE_CASE , out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) _lowercase : Union[str, Any] = UperNetConfig( backbone_config=SCREAMING_SNAKE_CASE , auxiliary_in_channels=SCREAMING_SNAKE_CASE , num_labels=SCREAMING_SNAKE_CASE , idalabel=SCREAMING_SNAKE_CASE , labelaid=SCREAMING_SNAKE_CASE , ) return config def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: _lowercase : Any = [] # fmt: off # stem rename_keys.append(('backbone.downsample_layers.0.0.weight', 'backbone.embeddings.patch_embeddings.weight') ) rename_keys.append(('backbone.downsample_layers.0.0.bias', 'backbone.embeddings.patch_embeddings.bias') ) rename_keys.append(('backbone.downsample_layers.0.1.weight', 'backbone.embeddings.layernorm.weight') ) rename_keys.append(('backbone.downsample_layers.0.1.bias', 'backbone.embeddings.layernorm.bias') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F"""backbone.stages.{i}.{j}.gamma""", F"""backbone.encoder.stages.{i}.layers.{j}.layer_scale_parameter""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.depthwise_conv.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.dwconv.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.depthwise_conv.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.dwconv.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.norm.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.layernorm.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.norm.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.layernorm.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv1.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv1.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv1.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv1.bias""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv2.weight""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv2.weight""") ) rename_keys.append((F"""backbone.stages.{i}.{j}.pointwise_conv2.bias""", F"""backbone.encoder.stages.{i}.layers.{j}.pwconv2.bias""") ) if i > 0: rename_keys.append((F"""backbone.downsample_layers.{i}.0.weight""", F"""backbone.encoder.stages.{i}.downsampling_layer.0.weight""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.0.bias""", F"""backbone.encoder.stages.{i}.downsampling_layer.0.bias""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.1.weight""", F"""backbone.encoder.stages.{i}.downsampling_layer.1.weight""") ) rename_keys.append((F"""backbone.downsample_layers.{i}.1.bias""", F"""backbone.encoder.stages.{i}.downsampling_layer.1.bias""") ) rename_keys.append((F"""backbone.norm{i}.weight""", F"""backbone.hidden_states_norms.stage{i+1}.weight""") ) rename_keys.append((F"""backbone.norm{i}.bias""", F"""backbone.hidden_states_norms.stage{i+1}.bias""") ) # decode head rename_keys.extend( [ ('decode_head.conv_seg.weight', 'decode_head.classifier.weight'), ('decode_head.conv_seg.bias', 'decode_head.classifier.bias'), ('auxiliary_head.conv_seg.weight', 'auxiliary_head.classifier.weight'), ('auxiliary_head.conv_seg.bias', 'auxiliary_head.classifier.bias'), ] ) # fmt: on return rename_keys def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Optional[int]: _lowercase : Any = dct.pop(SCREAMING_SNAKE_CASE ) _lowercase : Any = val def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Any: _lowercase : List[Any] = { 'upernet-convnext-tiny': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_tiny_fp16_512x512_160k_ade20k/upernet_convnext_tiny_fp16_512x512_160k_ade20k_20220227_124553-cad485de.pth', 'upernet-convnext-small': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_small_fp16_512x512_160k_ade20k/upernet_convnext_small_fp16_512x512_160k_ade20k_20220227_131208-1b1e394f.pth', 'upernet-convnext-base': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_base_fp16_512x512_160k_ade20k/upernet_convnext_base_fp16_512x512_160k_ade20k_20220227_181227-02a24fc6.pth', 'upernet-convnext-large': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_large_fp16_640x640_160k_ade20k/upernet_convnext_large_fp16_640x640_160k_ade20k_20220226_040532-e57aa54d.pth', 'upernet-convnext-xlarge': 'https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_xlarge_fp16_640x640_160k_ade20k/upernet_convnext_xlarge_fp16_640x640_160k_ade20k_20220226_080344-95fc38c2.pth', } _lowercase : Optional[int] = model_name_to_url[model_name] _lowercase : str = torch.hub.load_state_dict_from_url(SCREAMING_SNAKE_CASE , map_location='cpu' )['state_dict'] _lowercase : Optional[int] = get_upernet_config(SCREAMING_SNAKE_CASE ) _lowercase : Tuple = UperNetForSemanticSegmentation(SCREAMING_SNAKE_CASE ) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): _lowercase : List[Any] = state_dict.pop(SCREAMING_SNAKE_CASE ) if "bn" in key: _lowercase : Any = key.replace('bn' , 'batch_norm' ) _lowercase : Any = val # rename keys _lowercase : int = create_rename_keys(SCREAMING_SNAKE_CASE ) for src, dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) model.load_state_dict(SCREAMING_SNAKE_CASE ) # verify on image _lowercase : Union[str, Any] = 'https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg' _lowercase : Any = Image.open(requests.get(SCREAMING_SNAKE_CASE , stream=SCREAMING_SNAKE_CASE ).raw ).convert('RGB' ) _lowercase : Tuple = SegformerImageProcessor() _lowercase : Tuple = processor(SCREAMING_SNAKE_CASE , return_tensors='pt' ).pixel_values with torch.no_grad(): _lowercase : Dict = model(SCREAMING_SNAKE_CASE ) if model_name == "upernet-convnext-tiny": _lowercase : Dict = torch.tensor( [[-8.8110, -8.8110, -8.6521], [-8.8110, -8.8110, -8.6521], [-8.7746, -8.7746, -8.6130]] ) elif model_name == "upernet-convnext-small": _lowercase : Union[str, Any] = torch.tensor( [[-8.8236, -8.8236, -8.6771], [-8.8236, -8.8236, -8.6771], [-8.7638, -8.7638, -8.6240]] ) elif model_name == "upernet-convnext-base": _lowercase : Dict = torch.tensor( [[-8.8558, -8.8558, -8.6905], [-8.8558, -8.8558, -8.6905], [-8.7669, -8.7669, -8.6021]] ) elif model_name == "upernet-convnext-large": _lowercase : Optional[int] = torch.tensor( [[-8.6660, -8.6660, -8.6210], [-8.6660, -8.6660, -8.6210], [-8.6310, -8.6310, -8.5964]] ) elif model_name == "upernet-convnext-xlarge": _lowercase : str = torch.tensor( [[-8.4980, -8.4980, -8.3977], [-8.4980, -8.4980, -8.3977], [-8.4379, -8.4379, -8.3412]] ) print('Logits:' , outputs.logits[0, 0, :3, :3] ) assert torch.allclose(outputs.logits[0, 0, :3, :3] , SCREAMING_SNAKE_CASE , atol=1E-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(SCREAMING_SNAKE_CASE ) print(F"""Saving processor to {pytorch_dump_folder_path}""" ) processor.save_pretrained(SCREAMING_SNAKE_CASE ) if push_to_hub: print(F"""Pushing model and processor for {model_name} to hub""" ) model.push_to_hub(F"""openmmlab/{model_name}""" ) processor.push_to_hub(F"""openmmlab/{model_name}""" ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="upernet-convnext-tiny", type=str, choices=[f'''upernet-convnext-{size}''' for size in ["tiny", "small", "base", "large", "xlarge"]], help="Name of the ConvNext UperNet model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) UpperCamelCase = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
66
1
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class lowerCAmelCase_ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCamelCase : int = IFInpaintingSuperResolutionPipeline _UpperCamelCase : Any = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"width", "height"} _UpperCamelCase : Union[str, Any] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"original_image"} ) _UpperCamelCase : Tuple = PipelineTesterMixin.required_optional_params - {"latents"} def __a ( self ): return self._get_superresolution_dummy_components() def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : int = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Union[str, Any] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : Union[str, Any] = floats_tensor((1, 3, 1_6, 1_6) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : Optional[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'original_image': original_image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def __a ( self ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def __a ( self ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != 'cuda' , reason='float16 requires CUDA' ) def __a ( self ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __a ( self ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __a ( self ): self._test_save_load_local() def __a ( self ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
66
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto.configuration_auto import CONFIG_MAPPING UpperCamelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : Any = "upernet" def __init__( self , _lowerCAmelCase=None , _lowerCAmelCase=5_1_2 , _lowerCAmelCase=0.02 , _lowerCAmelCase=[1, 2, 3, 6] , _lowerCAmelCase=True , _lowerCAmelCase=0.4 , _lowerCAmelCase=3_8_4 , _lowerCAmelCase=2_5_6 , _lowerCAmelCase=1 , _lowerCAmelCase=False , _lowerCAmelCase=2_5_5 , **_lowerCAmelCase , ): super().__init__(**_lowerCAmelCase ) if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.' ) _lowercase : Optional[Any] = CONFIG_MAPPING['resnet'](out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): _lowercase : List[Any] = backbone_config.get('model_type' ) _lowercase : str = CONFIG_MAPPING[backbone_model_type] _lowercase : Tuple = config_class.from_dict(_lowerCAmelCase ) _lowercase : Optional[Any] = backbone_config _lowercase : Any = hidden_size _lowercase : Any = initializer_range _lowercase : Tuple = pool_scales _lowercase : List[Any] = use_auxiliary_head _lowercase : Optional[Any] = auxiliary_loss_weight _lowercase : Any = auxiliary_in_channels _lowercase : Any = auxiliary_channels _lowercase : List[str] = auxiliary_num_convs _lowercase : List[str] = auxiliary_concat_input _lowercase : Tuple = loss_ignore_index def __a ( self ): _lowercase : str = copy.deepcopy(self.__dict__ ) _lowercase : Tuple = self.backbone_config.to_dict() _lowercase : int = self.__class__.model_type return output
66
1
from typing import List import datasets from datasets.tasks import AudioClassification from ..folder_based_builder import folder_based_builder UpperCamelCase = datasets.utils.logging.get_logger(__name__) class lowerCAmelCase_ ( folder_based_builder.FolderBasedBuilderConfig ): _UpperCamelCase : bool = None _UpperCamelCase : bool = None class lowerCAmelCase_ ( folder_based_builder.FolderBasedBuilder ): _UpperCamelCase : Dict = datasets.Audio() _UpperCamelCase : str = "audio" _UpperCamelCase : Optional[Any] = AudioFolderConfig _UpperCamelCase : List[str] # definition at the bottom of the script _UpperCamelCase : Tuple = AudioClassification(audio_column="audio" , label_column="label" ) UpperCamelCase = [ ".aiff", ".au", ".avr", ".caf", ".flac", ".htk", ".svx", ".mat4", ".mat5", ".mpc2k", ".ogg", ".paf", ".pvf", ".raw", ".rf64", ".sd2", ".sds", ".ircam", ".voc", ".w64", ".wav", ".nist", ".wavex", ".wve", ".xi", ".mp3", ".opus", ] UpperCamelCase = AUDIO_EXTENSIONS
66
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: if a < 0 or b < 0: raise ValueError('the value of both inputs must be positive' ) _lowercase : str = str(bin(SCREAMING_SNAKE_CASE ) )[2:] # remove the leading "0b" _lowercase : int = str(bin(SCREAMING_SNAKE_CASE ) )[2:] # remove the leading "0b" _lowercase : List[Any] = max(len(SCREAMING_SNAKE_CASE ) , len(SCREAMING_SNAKE_CASE ) ) return "0b" + "".join( str(int(char_a == '1' and char_b == '1' ) ) for char_a, char_b in zip(a_binary.zfill(SCREAMING_SNAKE_CASE ) , b_binary.zfill(SCREAMING_SNAKE_CASE ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
66
1
from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLModel, ) class lowerCAmelCase_ : def __init__( self , _lowerCAmelCase , ): _lowercase : List[Any] = parent _lowercase : List[str] = 1_3 _lowercase : str = 7 _lowercase : int = 3_0 _lowercase : Optional[Any] = self.seq_length + self.mem_len _lowercase : List[str] = 1_5 _lowercase : Tuple = True _lowercase : Optional[Any] = True _lowercase : List[Any] = 9_9 _lowercase : int = [1_0, 5_0, 8_0] _lowercase : Tuple = 3_2 _lowercase : Any = 3_2 _lowercase : int = 4 _lowercase : List[str] = 8 _lowercase : Tuple = 1_2_8 _lowercase : int = 2 _lowercase : Optional[Any] = 2 _lowercase : str = None _lowercase : int = 1 _lowercase : Tuple = 0 _lowercase : Dict = 3 _lowercase : Union[str, Any] = self.vocab_size - 1 _lowercase : Optional[Any] = 0.01 def __a ( self ): _lowercase : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowercase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowercase : List[Any] = None if self.use_labels: _lowercase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowercase : Any = TransfoXLConfig( vocab_size=self.vocab_size , mem_len=self.mem_len , clamp_len=self.clamp_len , cutoffs=self.cutoffs , d_model=self.hidden_size , d_embed=self.d_embed , n_head=self.num_attention_heads , d_head=self.d_head , d_inner=self.d_inner , div_val=self.div_val , n_layer=self.num_hidden_layers , eos_token_id=self.eos_token_id , pad_token_id=self.vocab_size - 1 , init_range=self.init_range , num_labels=self.num_labels , ) return (config, input_ids_a, input_ids_a, lm_labels) def __a ( self ): random.seed(self.seed ) tf.random.set_seed(self.seed ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): _lowercase : Union[str, Any] = TFTransfoXLModel(_lowerCAmelCase ) _lowercase , _lowercase : Union[str, Any] = model(_lowerCAmelCase ).to_tuple() _lowercase : Any = {'input_ids': input_ids_a, 'mems': mems_a} _lowercase , _lowercase : Any = model(_lowerCAmelCase ).to_tuple() self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): _lowercase : Dict = TFTransfoXLLMHeadModel(_lowerCAmelCase ) _lowercase , _lowercase : List[Any] = model(_lowerCAmelCase ).to_tuple() _lowercase : Optional[int] = {'input_ids': input_ids_a, 'labels': lm_labels} _lowercase , _lowercase : Union[str, Any] = model(_lowerCAmelCase ).to_tuple() _lowercase , _lowercase : Dict = model([input_ids_a, mems_a] ).to_tuple() _lowercase : Any = {'input_ids': input_ids_a, 'mems': mems_a, 'labels': lm_labels} _lowercase , _lowercase : Union[str, Any] = model(_lowerCAmelCase ).to_tuple() self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): _lowercase : str = TFTransfoXLForSequenceClassification(_lowerCAmelCase ) _lowercase : Dict = model(_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __a ( self ): _lowercase : Any = self.prepare_config_and_inputs() ((_lowercase) , (_lowercase) , (_lowercase) , (_lowercase)) : str = config_and_inputs _lowercase : List[Any] = {'input_ids': input_ids_a} return config, inputs_dict @require_tf class lowerCAmelCase_ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCamelCase : List[str] = ( (TFTransfoXLModel, TFTransfoXLLMHeadModel, TFTransfoXLForSequenceClassification) if is_tf_available() else () ) _UpperCamelCase : Optional[Any] = () if is_tf_available() else () _UpperCamelCase : Any = ( { "feature-extraction": TFTransfoXLModel, "text-classification": TFTransfoXLForSequenceClassification, "text-generation": TFTransfoXLLMHeadModel, "zero-shot": TFTransfoXLForSequenceClassification, } if is_tf_available() else {} ) # TODO: add this test when TFTransfoXLLMHead has a linear output layer implemented _UpperCamelCase : int = False _UpperCamelCase : Tuple = False _UpperCamelCase : str = False _UpperCamelCase : Optional[Any] = False def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): if pipeline_test_casse_name == "TextGenerationPipelineTests": # Get `ValueError: AttributeError: 'NoneType' object has no attribute 'new_ones'` or `AssertionError`. # `TransfoXLConfig` was never used in pipeline tests: cannot create a simple # tokenizer. return True return False def __a ( self ): _lowercase : Optional[Any] = TFTransfoXLModelTester(self ) _lowercase : Union[str, Any] = ConfigTester(self , config_class=_lowerCAmelCase , d_embed=3_7 ) def __a ( self ): self.config_tester.run_common_tests() def __a ( self ): self.model_tester.set_seed() _lowercase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_model(*_lowerCAmelCase ) def __a ( self ): self.model_tester.set_seed() _lowercase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_lm_head(*_lowerCAmelCase ) def __a ( self ): _lowercase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_for_sequence_classification(*_lowerCAmelCase ) def __a ( self ): _lowercase , _lowercase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() _lowercase : Optional[Any] = [TFTransfoXLForSequenceClassification] for model_class in self.all_model_classes: _lowercase : int = model_class(_lowerCAmelCase ) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer ) if model_class in list_other_models_with_output_ebd: _lowercase : Dict = model.get_output_embeddings() assert isinstance(_lowerCAmelCase , tf.keras.layers.Layer ) _lowercase : List[Any] = model.get_bias() assert name is None else: _lowercase : List[Any] = model.get_output_embeddings() assert x is None _lowercase : int = model.get_bias() assert name is None def __a ( self ): # TODO JP: Make TransfoXL XLA compliant pass @slow def __a ( self ): for model_name in TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowercase : Union[str, Any] = TFTransfoXLModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) @unittest.skip(reason='This model doesn\'t play well with fit() due to not returning a single loss.' ) def __a ( self ): pass @require_tf class lowerCAmelCase_ ( unittest.TestCase ): @unittest.skip('Skip test until #12651 is resolved.' ) @slow def __a ( self ): _lowercase : str = TFTransfoXLLMHeadModel.from_pretrained('transfo-xl-wt103' ) # fmt: off _lowercase : Union[str, Any] = tf.convert_to_tensor([[3_3,1_2_9_7,2,1,1_0_0_9,4,1_1_0_9,1_1_7_3_9,4_7_6_2,3_5_8,5,2_5,2_4_5,2_2,1_7_0_6,1_7,2_0_0_9_8,5,3_2_1_5,2_1,3_7,1_1_1_0,3,1_3,1_0_4_1,4,2_4,6_0_3,4_9_0,2,7_1_4_7_7,2_0_0_9_8,1_0_4_4_4_7,2,2_0_9_6_1,1,2_6_0_4,4,1,3_2_9,3,6_2_2_4,8_3_1,1_6_0_0_2,2,8,6_0_3,7_8_9_6_7,2_9_5_4_6,2_3,8_0_3,2_0,2_5,4_1_6,5,8,2_3_2,4,2_7_7,6,1_8_5_5,4_6_0_1,3,2_9_5_4_6,5_4,8,3_6_0_9,5,5_7_2_1_1,4_9,4,1,2_7_7,1_8,8,1_7_5_5,1_5_6_9_1,3,3_4_1,2_5,4_1_6,6_9_3,4_2_5_7_3,7_1,1_7,4_0_1,9_4,3_1,1_7_9_1_9,2,2_9_5_4_6,7_8_7_3,1_8,1,4_3_5,2_3,1_1_0_1_1,7_5_5,5,5_1_6_7,3,7_9_8_3,9_8,8_4,2,2_9_5_4_6,3_2_6_7,8,3_6_0_9,4,1,4_8_6_5,1_0_7_5,2,6_0_8_7,7_1,6,3_4_6,8,5_8_5_4,3,2_9_5_4_6,8_2_4,1_4_0_0,1_8_6_8,2,1_9,1_6_0,2,3_1_1,8,5_4_9_6,2,2_0_9_2_0,1_7,2_5,1_5_0_9_7,3,2_4,2_4,0]] , dtype=tf.intaa ) # noqa: E231 # fmt: on # In 1991 , the remains of Russian Tsar Nicholas II and his family # ( except for Alexei and Maria ) are discovered . # The voice of Nicholas's young son , Tsarevich Alexei Nikolaevich , narrates the # remainder of the story . 1883 Western Siberia , # a young Grigori Rasputin is asked by his father and a group of men to perform magic . # Rasputin has a vision and denounces one of the men as a horse thief . Although his # father initially slaps him for making such an accusation , Rasputin watches as the # man is chased outside and beaten . Twenty years later , Rasputin sees a vision of # the Virgin Mary , prompting him to become a priest . Rasputin quickly becomes famous , # with people , even a bishop , begging for his blessing . <eod> </s> <eos> # fmt: off _lowercase : Optional[int] = [3_3,1_2_9_7,2,1,1_0_0_9,4,1_1_0_9,1_1_7_3_9,4_7_6_2,3_5_8,5,2_5,2_4_5,2_2,1_7_0_6,1_7,2_0_0_9_8,5,3_2_1_5,2_1,3_7,1_1_1_0,3,1_3,1_0_4_1,4,2_4,6_0_3,4_9_0,2,7_1_4_7_7,2_0_0_9_8,1_0_4_4_4_7,2,2_0_9_6_1,1,2_6_0_4,4,1,3_2_9,3,6_2_2_4,8_3_1,1_6_0_0_2,2,8,6_0_3,7_8_9_6_7,2_9_5_4_6,2_3,8_0_3,2_0,2_5,4_1_6,5,8,2_3_2,4,2_7_7,6,1_8_5_5,4_6_0_1,3,2_9_5_4_6,5_4,8,3_6_0_9,5,5_7_2_1_1,4_9,4,1,2_7_7,1_8,8,1_7_5_5,1_5_6_9_1,3,3_4_1,2_5,4_1_6,6_9_3,4_2_5_7_3,7_1,1_7,4_0_1,9_4,3_1,1_7_9_1_9,2,2_9_5_4_6,7_8_7_3,1_8,1,4_3_5,2_3,1_1_0_1_1,7_5_5,5,5_1_6_7,3,7_9_8_3,9_8,8_4,2,2_9_5_4_6,3_2_6_7,8,3_6_0_9,4,1,4_8_6_5,1_0_7_5,2,6_0_8_7,7_1,6,3_4_6,8,5_8_5_4,3,2_9_5_4_6,8_2_4,1_4_0_0,1_8_6_8,2,1_9,1_6_0,2,3_1_1,8,5_4_9_6,2,2_0_9_2_0,1_7,2_5,1_5_0_9_7,3,2_4,2_4,0,3_3,1,1_8_5_7,2,1,1_0_0_9,4,1_1_0_9,1_1_7_3_9,4_7_6_2,3_5_8,5,2_5,2_4_5,2_8,1_1_1_0,3,1_3,1_0_4_1,4,2_4,6_0_3,4_9_0,2,7_1_4_7_7,2_0_0_9_8,1_0_4_4_4_7,2,2_0_9_6_1,1,2_6_0_4,4,1,3_2_9,3,0] # noqa: E231 # fmt: on # In 1991, the remains of Russian Tsar Nicholas II and his family ( # except for Alexei and Maria ) are discovered. The voice of young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story. # 1883 Western Siberia, a young Grigori Rasputin is asked by his father # and a group of men to perform magic. Rasputin has a vision and # denounces one of the men as a horse thief. Although his father initially # slaps him for making such an accusation, Rasputin watches as the man # is chased outside and beaten. Twenty years later, Rasputin sees a vision # of the Virgin Mary, prompting him to become a priest. # Rasputin quickly becomes famous, with people, even a bishop, begging for # his blessing. <unk> <unk> <eos> In the 1990s, the remains of Russian Tsar # Nicholas II and his family were discovered. The voice of <unk> young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story.<eos> _lowercase : Union[str, Any] = model.generate(_lowerCAmelCase , max_length=2_0_0 , do_sample=_lowerCAmelCase ) self.assertListEqual(output_ids[0].numpy().tolist() , _lowerCAmelCase )
66
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class lowerCAmelCase_ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCamelCase : int = IFInpaintingSuperResolutionPipeline _UpperCamelCase : Any = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"width", "height"} _UpperCamelCase : Union[str, Any] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"original_image"} ) _UpperCamelCase : Tuple = PipelineTesterMixin.required_optional_params - {"latents"} def __a ( self ): return self._get_superresolution_dummy_components() def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : int = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Union[str, Any] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : Union[str, Any] = floats_tensor((1, 3, 1_6, 1_6) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : Optional[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'original_image': original_image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def __a ( self ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def __a ( self ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != 'cuda' , reason='float16 requires CUDA' ) def __a ( self ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def __a ( self ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def __a ( self ): self._test_save_load_local() def __a ( self ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
66
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCamelCase = { "configuration_squeezebert": [ "SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "SqueezeBertConfig", "SqueezeBertOnnxConfig", ], "tokenization_squeezebert": ["SqueezeBertTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["SqueezeBertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "SqueezeBertForMaskedLM", "SqueezeBertForMultipleChoice", "SqueezeBertForQuestionAnswering", "SqueezeBertForSequenceClassification", "SqueezeBertForTokenClassification", "SqueezeBertModel", "SqueezeBertModule", "SqueezeBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_squeezebert import ( SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, SqueezeBertConfig, SqueezeBertOnnxConfig, ) from .tokenization_squeezebert import SqueezeBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_squeezebert_fast import SqueezeBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_squeezebert import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, SqueezeBertModule, SqueezeBertPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> tuple[float, float]: # Check if the input is valid if not len(SCREAMING_SNAKE_CASE ) == len(SCREAMING_SNAKE_CASE ) == 3: raise ValueError('Please enter a valid equation.' ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError('Both a & b of two equations can\'t be zero.' ) # Extract the coefficients _lowercase , _lowercase , _lowercase : Tuple = equationa _lowercase , _lowercase , _lowercase : Dict = equationa # Calculate the determinants of the matrices _lowercase : str = aa * ba - aa * ba _lowercase : Any = ca * ba - ca * ba _lowercase : Optional[int] = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError('Infinite solutions. (Consistent system)' ) else: raise ValueError('No solution. (Inconsistent system)' ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: _lowercase : Union[str, Any] = determinant_x / determinant _lowercase : Tuple = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
66
1
# DISCLAIMER: This code is strongly influenced by https://github.com/pesser/pytorch_diffusion # and https://github.com/hojonathanho/diffusion import math from dataclasses import dataclass from typing import List, Optional, Tuple, Union import numpy as np import torch from diffusers.configuration_utils import ConfigMixin, register_to_config from diffusers.schedulers.scheduling_utils import SchedulerMixin from diffusers.utils import BaseOutput, deprecate @dataclass # Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->DDIM class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : torch.FloatTensor _UpperCamelCase : Optional[torch.FloatTensor] = None def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=0.999 , SCREAMING_SNAKE_CASE="cosine" , ) -> Tuple: if alpha_transform_type == "cosine": def alpha_bar_fn(SCREAMING_SNAKE_CASE ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(SCREAMING_SNAKE_CASE ): return math.exp(t * -12.0 ) else: raise ValueError(F"""Unsupported alpha_tranform_type: {alpha_transform_type}""" ) _lowercase : Dict = [] for i in range(SCREAMING_SNAKE_CASE ): _lowercase : str = i / num_diffusion_timesteps _lowercase : int = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(SCREAMING_SNAKE_CASE ) / alpha_bar_fn(SCREAMING_SNAKE_CASE ) , SCREAMING_SNAKE_CASE ) ) return torch.tensor(SCREAMING_SNAKE_CASE , dtype=torch.floataa ) class lowerCAmelCase_ ( __snake_case , __snake_case ): _UpperCamelCase : Optional[int] = 1 @register_to_config def __init__( self , _lowerCAmelCase = 1_0_0_0 , _lowerCAmelCase = 0.00_01 , _lowerCAmelCase = 0.02 , _lowerCAmelCase = "linear" , _lowerCAmelCase = None , _lowerCAmelCase = True , _lowerCAmelCase = True , _lowerCAmelCase = 0 , _lowerCAmelCase = "epsilon" , _lowerCAmelCase = 1.0 , **_lowerCAmelCase , ): if kwargs.get('set_alpha_to_one' , _lowerCAmelCase ) is not None: _lowercase : Optional[int] = ( 'The `set_alpha_to_one` argument is deprecated. Please use `set_alpha_to_zero` instead.' ) deprecate('set_alpha_to_one' , '1.0.0' , _lowerCAmelCase , standard_warn=_lowerCAmelCase ) _lowercase : str = kwargs['set_alpha_to_one'] if trained_betas is not None: _lowercase : List[str] = torch.tensor(_lowerCAmelCase , dtype=torch.floataa ) elif beta_schedule == "linear": _lowercase : str = torch.linspace(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , dtype=torch.floataa ) elif beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. _lowercase : Optional[int] = ( torch.linspace(beta_start**0.5 , beta_end**0.5 , _lowerCAmelCase , dtype=torch.floataa ) ** 2 ) elif beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule _lowercase : str = betas_for_alpha_bar(_lowerCAmelCase ) else: raise NotImplementedError(F"""{beta_schedule} does is not implemented for {self.__class__}""" ) _lowercase : Dict = 1.0 - self.betas _lowercase : Tuple = torch.cumprod(self.alphas , dim=0 ) # At every step in inverted ddim, we are looking into the next alphas_cumprod # For the final step, there is no next alphas_cumprod, and the index is out of bounds # `set_alpha_to_zero` decides whether we set this parameter simply to zero # in this case, self.step() just output the predicted noise # or whether we use the final alpha of the "non-previous" one. _lowercase : Union[str, Any] = torch.tensor(0.0 ) if set_alpha_to_zero else self.alphas_cumprod[-1] # standard deviation of the initial noise distribution _lowercase : Tuple = 1.0 # setable values _lowercase : List[str] = None _lowercase : int = torch.from_numpy(np.arange(0 , _lowerCAmelCase ).copy().astype(np.intaa ) ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase = None ): return sample def __a ( self , _lowerCAmelCase , _lowerCAmelCase = None ): if num_inference_steps > self.config.num_train_timesteps: raise ValueError( F"""`num_inference_steps`: {num_inference_steps} cannot be larger than `self.config.train_timesteps`:""" F""" {self.config.num_train_timesteps} as the unet model trained with this scheduler can only handle""" F""" maximal {self.config.num_train_timesteps} timesteps.""" ) _lowercase : int = num_inference_steps _lowercase : int = self.config.num_train_timesteps // self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 _lowercase : Optional[int] = (np.arange(0 , _lowerCAmelCase ) * step_ratio).round().copy().astype(np.intaa ) _lowercase : Optional[int] = torch.from_numpy(_lowerCAmelCase ).to(_lowerCAmelCase ) self.timesteps += self.config.steps_offset def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = 0.0 , _lowerCAmelCase = False , _lowerCAmelCase = None , _lowerCAmelCase = True , ): # 1. get previous step value (=t+1) _lowercase : str = timestep + self.config.num_train_timesteps // self.num_inference_steps # 2. compute alphas, betas # change original implementation to exactly match noise levels for analogous forward process _lowercase : List[str] = self.alphas_cumprod[timestep] _lowercase : Tuple = ( self.alphas_cumprod[prev_timestep] if prev_timestep < self.config.num_train_timesteps else self.final_alpha_cumprod ) _lowercase : Any = 1 - alpha_prod_t # 3. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf if self.config.prediction_type == "epsilon": _lowercase : Optional[Any] = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 _lowercase : List[Any] = model_output elif self.config.prediction_type == "sample": _lowercase : Optional[int] = model_output _lowercase : Any = (sample - alpha_prod_t ** 0.5 * pred_original_sample) / beta_prod_t ** 0.5 elif self.config.prediction_type == "v_prediction": _lowercase : Any = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output _lowercase : List[Any] = (alpha_prod_t**0.5) * model_output + (beta_prod_t**0.5) * sample else: raise ValueError( F"""prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample`, or""" ' `v_prediction`' ) # 4. Clip or threshold "predicted x_0" if self.config.clip_sample: _lowercase : Optional[Any] = pred_original_sample.clamp( -self.config.clip_sample_range , self.config.clip_sample_range ) # 5. compute "direction pointing to x_t" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _lowercase : str = (1 - alpha_prod_t_prev) ** 0.5 * pred_epsilon # 6. compute x_t without "random noise" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _lowercase : Optional[int] = alpha_prod_t_prev ** 0.5 * pred_original_sample + pred_sample_direction if not return_dict: return (prev_sample, pred_original_sample) return DDIMSchedulerOutput(prev_sample=_lowerCAmelCase , pred_original_sample=_lowerCAmelCase ) def __len__( self ): return self.config.num_train_timesteps
66
def __magic_name__ ( SCREAMING_SNAKE_CASE = 50 ) -> int: _lowercase : Optional[int] = [[0] * 3 for _ in range(length + 1 )] for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): different_colour_ways_number[row_length][tile_length - 2] += ( different_colour_ways_number[row_length - tile_start - tile_length][ tile_length - 2 ] + 1 ) return sum(different_colour_ways_number[length] ) if __name__ == "__main__": print(f'''{solution() = }''')
66
1
import unittest from transformers import BertGenerationTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin UpperCamelCase = "▁" UpperCamelCase = get_tests_dir("fixtures/test_sentencepiece.model") @require_sentencepiece class lowerCAmelCase_ ( __snake_case , unittest.TestCase ): _UpperCamelCase : int = BertGenerationTokenizer _UpperCamelCase : List[str] = False _UpperCamelCase : Union[str, Any] = True def __a ( self ): super().setUp() _lowercase : Union[str, Any] = BertGenerationTokenizer(_lowerCAmelCase , keep_accents=_lowerCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def __a ( self ): _lowercase : str = '<s>' _lowercase : Dict = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_lowerCAmelCase ) , _lowerCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_lowerCAmelCase ) , _lowerCAmelCase ) def __a ( self ): _lowercase : Optional[int] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<unk>' ) self.assertEqual(vocab_keys[1] , '<s>' ) self.assertEqual(vocab_keys[-1] , '<pad>' ) self.assertEqual(len(_lowerCAmelCase ) , 1_0_0_2 ) def __a ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1_0_0_0 ) def __a ( self ): _lowercase : int = BertGenerationTokenizer(_lowerCAmelCase , keep_accents=_lowerCAmelCase ) _lowercase : Tuple = tokenizer.tokenize('This is a test' ) self.assertListEqual(_lowerCAmelCase , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) , [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2] , ) _lowercase : str = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) _lowercase : int = tokenizer.convert_tokens_to_ids(_lowerCAmelCase ) self.assertListEqual( _lowerCAmelCase , [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 0, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 0, 4] , ) _lowercase : Optional[Any] = tokenizer.convert_ids_to_tokens(_lowerCAmelCase ) self.assertListEqual( _lowerCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) @cached_property def __a ( self ): return BertGenerationTokenizer.from_pretrained('google/bert_for_seq_generation_L-24_bbc_encoder' ) @slow def __a ( self ): _lowercase : Union[str, Any] = 'Hello World!' _lowercase : Dict = [1_8_5_3_6, 2_2_6_0, 1_0_1] self.assertListEqual(_lowerCAmelCase , self.big_tokenizer.encode(_lowerCAmelCase ) ) @slow def __a ( self ): _lowercase : Any = ( 'This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will' ' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth' ) _lowercase : int = [ 8_7_1, 4_1_9, 3_5_8, 9_4_6, 9_9_1, 2_5_2_1, 4_5_2, 3_5_8, 1_3_5_7, 3_8_7, 7_7_5_1, 3_5_3_6, 1_1_2, 9_8_5, 4_5_6, 1_2_6, 8_6_5, 9_3_8, 5_4_0_0, 5_7_3_4, 4_5_8, 1_3_6_8, 4_6_7, 7_8_6, 2_4_6_2, 5_2_4_6, 1_1_5_9, 6_3_3, 8_6_5, 4_5_1_9, 4_5_7, 5_8_2, 8_5_2, 2_5_5_7, 4_2_7, 9_1_6, 5_0_8, 4_0_5, 3_4_3_2_4, 4_9_7, 3_9_1, 4_0_8, 1_1_3_4_2, 1_2_4_4, 3_8_5, 1_0_0, 9_3_8, 9_8_5, 4_5_6, 5_7_4, 3_6_2, 1_2_5_9_7, 3_2_0_0, 3_1_2_9, 1_1_7_2, ] self.assertListEqual(_lowerCAmelCase , self.big_tokenizer.encode(_lowerCAmelCase ) ) @require_torch @slow def __a ( self ): import torch from transformers import BertGenerationConfig, BertGenerationEncoder # Build sequence _lowercase : List[str] = list(self.big_tokenizer.get_vocab().keys() )[:1_0] _lowercase : Union[str, Any] = ' '.join(_lowerCAmelCase ) _lowercase : int = self.big_tokenizer.encode_plus(_lowerCAmelCase , return_tensors='pt' , return_token_type_ids=_lowerCAmelCase ) _lowercase : Any = self.big_tokenizer.batch_encode_plus( [sequence + ' ' + sequence] , return_tensors='pt' , return_token_type_ids=_lowerCAmelCase ) _lowercase : Optional[int] = BertGenerationConfig() _lowercase : Any = BertGenerationEncoder(_lowerCAmelCase ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**_lowerCAmelCase ) model(**_lowerCAmelCase ) @slow def __a ( self ): # fmt: off _lowercase : List[Any] = {'input_ids': [[3_9_2_8_6, 4_5_8, 3_6_3_3_5, 2_0_0_1, 4_5_6, 1_3_0_7_3, 1_3_2_6_6, 4_5_5, 1_1_3, 7_7_4_6, 1_7_4_1, 1_1_1_5_7, 3_9_1, 1_3_0_7_3, 1_3_2_6_6, 4_5_5, 1_1_3, 3_9_6_7, 3_5_4_1_2, 1_1_3, 4_9_3_6, 1_0_9, 3_8_7_0, 2_3_7_7, 1_1_3, 3_0_0_8_4, 4_5_7_2_0, 4_5_8, 1_3_4, 1_7_4_9_6, 1_1_2, 5_0_3, 1_1_6_7_2, 1_1_3, 1_1_8, 1_1_2, 5_6_6_5, 1_3_3_4_7, 3_8_6_8_7, 1_1_2, 1_4_9_6, 3_1_3_8_9, 1_1_2, 3_2_6_8, 4_7_2_6_4, 1_3_4, 9_6_2, 1_1_2, 1_6_3_7_7, 8_0_3_5, 2_3_1_3_0, 4_3_0, 1_2_1_6_9, 1_5_5_1_8, 2_8_5_9_2, 4_5_8, 1_4_6, 4_1_6_9_7, 1_0_9, 3_9_1, 1_2_1_6_9, 1_5_5_1_8, 1_6_6_8_9, 4_5_8, 1_4_6, 4_1_3_5_8, 1_0_9, 4_5_2, 7_2_6, 4_0_3_4, 1_1_1, 7_6_3, 3_5_4_1_2, 5_0_8_2, 3_8_8, 1_9_0_3, 1_1_1, 9_0_5_1, 3_9_1, 2_8_7_0, 4_8_9_1_8, 1_9_0_0, 1_1_2_3, 5_5_0, 9_9_8, 1_1_2, 9_5_8_6, 1_5_9_8_5, 4_5_5, 3_9_1, 4_1_0, 2_2_9_5_5, 3_7_6_3_6, 1_1_4], [4_4_8, 1_7_4_9_6, 4_1_9, 3_6_6_3, 3_8_5, 7_6_3, 1_1_3, 2_7_5_3_3, 2_8_7_0, 3_2_8_3, 1_3_0_4_3, 1_6_3_9, 2_4_7_1_3, 5_2_3, 6_5_6, 2_4_0_1_3, 1_8_5_5_0, 2_5_2_1, 5_1_7, 2_7_0_1_4, 2_1_2_4_4, 4_2_0, 1_2_1_2, 1_4_6_5, 3_9_1, 9_2_7, 4_8_3_3, 3_8_8, 5_7_8, 1_1_7_8_6, 1_1_4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [4_8_4, 2_1_6_9, 7_6_8_7, 2_1_9_3_2, 1_8_1_4_6, 7_2_6, 3_6_3, 1_7_0_3_2, 3_3_9_1, 1_1_4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_lowerCAmelCase , model_name='google/bert_for_seq_generation_L-24_bbc_encoder' , revision='c817d1fd1be2ffa69431227a1fe320544943d4db' , )
66
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) UpperCamelCase = { "configuration_convnext": ["CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP", "ConvNextConfig", "ConvNextOnnxConfig"] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["ConvNextFeatureExtractor"] UpperCamelCase = ["ConvNextImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST", "ConvNextForImageClassification", "ConvNextModel", "ConvNextPreTrainedModel", "ConvNextBackbone", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "TFConvNextForImageClassification", "TFConvNextModel", "TFConvNextPreTrainedModel", ] if TYPE_CHECKING: from .configuration_convnext import CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvNextConfig, ConvNextOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_convnext import ConvNextFeatureExtractor from .image_processing_convnext import ConvNextImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convnext import ( CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvNextBackbone, ConvNextForImageClassification, ConvNextModel, ConvNextPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convnext import TFConvNextForImageClassification, TFConvNextModel, TFConvNextPreTrainedModel else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure)
66
1
from functools import lru_cache def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> set: _lowercase : Union[str, Any] = 2 _lowercase : Optional[Any] = set() while i * i <= n: if n % i: i += 1 else: n //= i factors.add(SCREAMING_SNAKE_CASE ) if n > 1: factors.add(SCREAMING_SNAKE_CASE ) return factors @lru_cache def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> int: return len(unique_prime_factors(SCREAMING_SNAKE_CASE ) ) def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> bool: return len(set(SCREAMING_SNAKE_CASE ) ) in (0, 1) def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> list: _lowercase : Tuple = 2 while True: # Increment each value of a generated range _lowercase : List[str] = [base + i for i in range(SCREAMING_SNAKE_CASE )] # Run elements through out unique_prime_factors function # Append our target number to the end. _lowercase : List[str] = [upf_len(SCREAMING_SNAKE_CASE ) for x in group] checker.append(SCREAMING_SNAKE_CASE ) # If all numbers in the list are equal, return the group variable. if equality(SCREAMING_SNAKE_CASE ): return group # Increment our base variable by 1 base += 1 def __magic_name__ ( SCREAMING_SNAKE_CASE = 4 ) -> int: _lowercase : Dict = run(SCREAMING_SNAKE_CASE ) return results[0] if len(SCREAMING_SNAKE_CASE ) else None if __name__ == "__main__": print(solution())
66
def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> bool: if p < 2: raise ValueError('p should not be less than 2!' ) elif p == 2: return True _lowercase : Optional[Any] = 4 _lowercase : Tuple = (1 << p) - 1 for _ in range(p - 2 ): _lowercase : Union[str, Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
66
1
import json import os from functools import lru_cache from typing import List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = {"vocab_file": "vocab.json", "merges_file": "merges.txt"} UpperCamelCase = { "vocab_file": { "allenai/longformer-base-4096": "https://huggingface.co/allenai/longformer-base-4096/resolve/main/vocab.json", "allenai/longformer-large-4096": ( "https://huggingface.co/allenai/longformer-large-4096/resolve/main/vocab.json" ), "allenai/longformer-large-4096-finetuned-triviaqa": ( "https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/vocab.json" ), "allenai/longformer-base-4096-extra.pos.embd.only": ( "https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/vocab.json" ), "allenai/longformer-large-4096-extra.pos.embd.only": ( "https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/vocab.json" ), }, "merges_file": { "allenai/longformer-base-4096": "https://huggingface.co/allenai/longformer-base-4096/resolve/main/merges.txt", "allenai/longformer-large-4096": ( "https://huggingface.co/allenai/longformer-large-4096/resolve/main/merges.txt" ), "allenai/longformer-large-4096-finetuned-triviaqa": ( "https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/merges.txt" ), "allenai/longformer-base-4096-extra.pos.embd.only": ( "https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/merges.txt" ), "allenai/longformer-large-4096-extra.pos.embd.only": ( "https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/merges.txt" ), }, } UpperCamelCase = { "allenai/longformer-base-4096": 4_096, "allenai/longformer-large-4096": 4_096, "allenai/longformer-large-4096-finetuned-triviaqa": 4_096, "allenai/longformer-base-4096-extra.pos.embd.only": 4_096, "allenai/longformer-large-4096-extra.pos.embd.only": 4_096, } @lru_cache() # Copied from transformers.models.roberta.tokenization_roberta.bytes_to_unicode def __magic_name__ ( ) -> Tuple: _lowercase : List[Any] = ( list(range(ord('!' ) , ord('~' ) + 1 ) ) + list(range(ord('¡' ) , ord('¬' ) + 1 ) ) + list(range(ord('®' ) , ord('ÿ' ) + 1 ) ) ) _lowercase : str = bs[:] _lowercase : Optional[Any] = 0 for b in range(2**8 ): if b not in bs: bs.append(SCREAMING_SNAKE_CASE ) cs.append(2**8 + n ) n += 1 _lowercase : str = [chr(SCREAMING_SNAKE_CASE ) for n in cs] return dict(zip(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> Any: _lowercase : int = set() _lowercase : Union[str, Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) _lowercase : str = char return pairs class lowerCAmelCase_ ( __snake_case ): _UpperCamelCase : List[Any] = VOCAB_FILES_NAMES _UpperCamelCase : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP _UpperCamelCase : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCamelCase : Dict = ["input_ids", "attention_mask"] def __init__( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase="replace" , _lowerCAmelCase="<s>" , _lowerCAmelCase="</s>" , _lowerCAmelCase="</s>" , _lowerCAmelCase="<s>" , _lowerCAmelCase="<unk>" , _lowerCAmelCase="<pad>" , _lowerCAmelCase="<mask>" , _lowerCAmelCase=False , **_lowerCAmelCase , ): _lowercase : List[str] = AddedToken(_lowerCAmelCase , lstrip=_lowerCAmelCase , rstrip=_lowerCAmelCase ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ) else bos_token _lowercase : str = AddedToken(_lowerCAmelCase , lstrip=_lowerCAmelCase , rstrip=_lowerCAmelCase ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ) else eos_token _lowercase : int = AddedToken(_lowerCAmelCase , lstrip=_lowerCAmelCase , rstrip=_lowerCAmelCase ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ) else sep_token _lowercase : Union[str, Any] = AddedToken(_lowerCAmelCase , lstrip=_lowerCAmelCase , rstrip=_lowerCAmelCase ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ) else cls_token _lowercase : Optional[int] = AddedToken(_lowerCAmelCase , lstrip=_lowerCAmelCase , rstrip=_lowerCAmelCase ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ) else unk_token _lowercase : str = AddedToken(_lowerCAmelCase , lstrip=_lowerCAmelCase , rstrip=_lowerCAmelCase ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ) else pad_token # Mask token behave like a normal word, i.e. include the space before it _lowercase : Dict = AddedToken(_lowerCAmelCase , lstrip=_lowerCAmelCase , rstrip=_lowerCAmelCase ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ) else mask_token super().__init__( errors=_lowerCAmelCase , bos_token=_lowerCAmelCase , eos_token=_lowerCAmelCase , unk_token=_lowerCAmelCase , sep_token=_lowerCAmelCase , cls_token=_lowerCAmelCase , pad_token=_lowerCAmelCase , mask_token=_lowerCAmelCase , add_prefix_space=_lowerCAmelCase , **_lowerCAmelCase , ) with open(_lowerCAmelCase , encoding='utf-8' ) as vocab_handle: _lowercase : Optional[int] = json.load(_lowerCAmelCase ) _lowercase : int = {v: k for k, v in self.encoder.items()} _lowercase : str = errors # how to handle errors in decoding _lowercase : int = bytes_to_unicode() _lowercase : Dict = {v: k for k, v in self.byte_encoder.items()} with open(_lowerCAmelCase , encoding='utf-8' ) as merges_handle: _lowercase : Dict = merges_handle.read().split('\n' )[1:-1] _lowercase : Tuple = [tuple(merge.split() ) for merge in bpe_merges] _lowercase : Union[str, Any] = dict(zip(_lowerCAmelCase , range(len(_lowerCAmelCase ) ) ) ) _lowercase : Dict = {} _lowercase : Optional[int] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions _lowercase : Optional[Any] = re.compile(r'\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+' ) @property def __a ( self ): return len(self.encoder ) def __a ( self ): return dict(self.encoder , **self.added_tokens_encoder ) def __a ( self , _lowerCAmelCase ): if token in self.cache: return self.cache[token] _lowercase : Any = tuple(_lowerCAmelCase ) _lowercase : List[str] = get_pairs(_lowerCAmelCase ) if not pairs: return token while True: _lowercase : Union[str, Any] = min(_lowerCAmelCase , key=lambda _lowerCAmelCase : self.bpe_ranks.get(_lowerCAmelCase , float('inf' ) ) ) if bigram not in self.bpe_ranks: break _lowercase , _lowercase : Dict = bigram _lowercase : int = [] _lowercase : Union[str, Any] = 0 while i < len(_lowerCAmelCase ): try: _lowercase : Optional[Any] = word.index(_lowerCAmelCase , _lowerCAmelCase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) _lowercase : List[str] = j if word[i] == first and i < len(_lowerCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 _lowercase : List[str] = tuple(_lowerCAmelCase ) _lowercase : Any = new_word if len(_lowerCAmelCase ) == 1: break else: _lowercase : List[str] = get_pairs(_lowerCAmelCase ) _lowercase : Tuple = ' '.join(_lowerCAmelCase ) _lowercase : Any = word return word def __a ( self , _lowerCAmelCase ): _lowercase : Tuple = [] for token in re.findall(self.pat , _lowerCAmelCase ): _lowercase : Optional[int] = ''.join( self.byte_encoder[b] for b in token.encode('utf-8' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(_lowerCAmelCase ).split(' ' ) ) return bpe_tokens def __a ( self , _lowerCAmelCase ): return self.encoder.get(_lowerCAmelCase , self.encoder.get(self.unk_token ) ) def __a ( self , _lowerCAmelCase ): return self.decoder.get(_lowerCAmelCase ) def __a ( self , _lowerCAmelCase ): _lowercase : List[Any] = ''.join(_lowerCAmelCase ) _lowercase : Dict = bytearray([self.byte_decoder[c] for c in text] ).decode('utf-8' , errors=self.errors ) return text def __a ( self , _lowerCAmelCase , _lowerCAmelCase = None ): if not os.path.isdir(_lowerCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _lowercase : List[Any] = os.path.join( _lowerCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) _lowercase : Optional[Any] = os.path.join( _lowerCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) with open(_lowerCAmelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_lowerCAmelCase , ensure_ascii=_lowerCAmelCase ) + '\n' ) _lowercase : str = 0 with open(_lowerCAmelCase , 'w' , encoding='utf-8' ) as writer: writer.write('#version: 0.2\n' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _lowerCAmelCase : kv[1] ): if index != token_index: logger.warning( F"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ' Please check that the tokenizer is not corrupted!' ) _lowercase : int = token_index writer.write(' '.join(_lowerCAmelCase ) + '\n' ) index += 1 return vocab_file, merge_file def __a ( self , _lowerCAmelCase , _lowerCAmelCase = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _lowercase : Optional[int] = [self.cls_token_id] _lowercase : Any = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __a ( self , _lowerCAmelCase , _lowerCAmelCase = None , _lowerCAmelCase = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_lowerCAmelCase , token_ids_a=_lowerCAmelCase , already_has_special_tokens=_lowerCAmelCase ) if token_ids_a is None: return [1] + ([0] * len(_lowerCAmelCase )) + [1] return [1] + ([0] * len(_lowerCAmelCase )) + [1, 1] + ([0] * len(_lowerCAmelCase )) + [1] def __a ( self , _lowerCAmelCase , _lowerCAmelCase = None ): _lowercase : Any = [self.sep_token_id] _lowercase : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __a ( self , _lowerCAmelCase , _lowerCAmelCase=False , **_lowerCAmelCase ): _lowercase : Optional[int] = kwargs.pop('add_prefix_space' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(_lowerCAmelCase ) > 0 and not text[0].isspace()): _lowercase : Dict = ' ' + text return (text, kwargs)
66
import random def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = False ) -> dict: _lowercase : dict = {i: [] for i in range(SCREAMING_SNAKE_CASE )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(SCREAMING_SNAKE_CASE ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(SCREAMING_SNAKE_CASE ): for j in range(i + 1 , SCREAMING_SNAKE_CASE ): if random.random() < probability: graph[i].append(SCREAMING_SNAKE_CASE ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(SCREAMING_SNAKE_CASE ) return graph def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> dict: return { i: [j for j in range(SCREAMING_SNAKE_CASE ) if i != j] for i in range(SCREAMING_SNAKE_CASE ) } if __name__ == "__main__": import doctest doctest.testmod()
66
1
import sacrebleu as scb from packaging import version from sacrebleu import TER import datasets UpperCamelCase = "\\n@inproceedings{snover-etal-2006-study,\n title = \"A Study of Translation Edit Rate with Targeted Human Annotation\",\n author = \"Snover, Matthew and\n Dorr, Bonnie and\n Schwartz, Rich and\n Micciulla, Linnea and\n Makhoul, John\",\n booktitle = \"Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers\",\n month = aug # \" 8-12\",\n year = \"2006\",\n address = \"Cambridge, Massachusetts, USA\",\n publisher = \"Association for Machine Translation in the Americas\",\n url = \"https://aclanthology.org/2006.amta-papers.25\",\n pages = \"223--231\",\n}\n@inproceedings{post-2018-call,\n title = \"A Call for Clarity in Reporting {BLEU} Scores\",\n author = \"Post, Matt\",\n booktitle = \"Proceedings of the Third Conference on Machine Translation: Research Papers\",\n month = oct,\n year = \"2018\",\n address = \"Belgium, Brussels\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/W18-6319\",\n pages = \"186--191\",\n}\n" UpperCamelCase = "\\nTER (Translation Edit Rate, also called Translation Error Rate) is a metric to quantify the edit operations that a\nhypothesis requires to match a reference translation. We use the implementation that is already present in sacrebleu\n(https://github.com/mjpost/sacreBLEU#ter), which in turn is inspired by the TERCOM implementation, which can be found\nhere: https://github.com/jhclark/tercom.\n\nThe implementation here is slightly different from sacrebleu in terms of the required input format. The length of\nthe references and hypotheses lists need to be the same, so you may need to transpose your references compared to\nsacrebleu's required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534\n\nSee the README.md file at https://github.com/mjpost/sacreBLEU#ter for more information.\n" UpperCamelCase = "\nProduces TER scores alongside the number of edits and reference length.\n\nArgs:\n predictions (list of str): The system stream (a sequence of segments).\n references (list of list of str): A list of one or more reference streams (each a sequence of segments).\n normalized (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`.\n ignore_punct (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`.\n support_zh_ja_chars (boolean): If `True`, tokenization/normalization supports processing of Chinese characters,\n as well as Japanese Kanji, Hiragana, Katakana, and Phonetic Extensions of Katakana.\n Only applies if `normalized = True`. Defaults to `False`.\n case_sensitive (boolean): If `False`, makes all predictions and references lowercase to ignore differences in case. Defaults to `False`.\n\nReturns:\n 'score' (float): TER score (num_edits / sum_ref_lengths * 100)\n 'num_edits' (int): The cumulative number of edits\n 'ref_length' (float): The cumulative average reference length\n\nExamples:\n Example 1:\n >>> predictions = [\"does this sentence match??\",\n ... \"what about this sentence?\",\n ... \"What did the TER metric user say to the developer?\"]\n >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"],\n ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"],\n ... [\"Your jokes are...\", \"...TERrible\"]]\n >>> ter = datasets.load_metric(\"ter\")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... case_sensitive=True)\n >>> print(results)\n {'score': 150.0, 'num_edits': 15, 'ref_length': 10.0}\n\n Example 2:\n >>> predictions = [\"does this sentence match??\",\n ... \"what about this sentence?\"]\n >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"],\n ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"]]\n >>> ter = datasets.load_metric(\"ter\")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... case_sensitive=True)\n >>> print(results)\n {'score': 62.5, 'num_edits': 5, 'ref_length': 8.0}\n\n Example 3:\n >>> predictions = [\"does this sentence match??\",\n ... \"what about this sentence?\"]\n >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"],\n ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"]]\n >>> ter = datasets.load_metric(\"ter\")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... normalized=True,\n ... case_sensitive=True)\n >>> print(results)\n {'score': 57.14285714285714, 'num_edits': 6, 'ref_length': 10.5}\n\n Example 4:\n >>> predictions = [\"does this sentence match??\",\n ... \"what about this sentence?\"]\n >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"],\n ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"]]\n >>> ter = datasets.load_metric(\"ter\")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... ignore_punct=True,\n ... case_sensitive=False)\n >>> print(results)\n {'score': 0.0, 'num_edits': 0, 'ref_length': 8.0}\n\n Example 5:\n >>> predictions = [\"does this sentence match??\",\n ... \"what about this sentence?\",\n ... \"What did the TER metric user say to the developer?\"]\n >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"],\n ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"],\n ... [\"Your jokes are...\", \"...TERrible\"]]\n >>> ter = datasets.load_metric(\"ter\")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... ignore_punct=True,\n ... case_sensitive=False)\n >>> print(results)\n {'score': 100.0, 'num_edits': 10, 'ref_length': 10.0}\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase_ ( datasets.Metric ): def __a ( self ): if version.parse(scb.__version__ ) < version.parse('1.4.12' ): raise ImportWarning( 'To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn\'t match this condition.\n' 'You can install it with `pip install "sacrebleu>=1.4.12"`.' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='http://www.cs.umd.edu/~snover/tercom/' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Sequence(datasets.Value('string' , id='sequence' ) , id='references' ), } ) , codebase_urls=['https://github.com/mjpost/sacreBLEU#ter'] , reference_urls=[ 'https://github.com/jhclark/tercom', ] , ) def __a ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = False , _lowerCAmelCase = False , _lowerCAmelCase = False , _lowerCAmelCase = False , ): _lowercase : Optional[int] = len(references[0] ) if any(len(_lowerCAmelCase ) != references_per_prediction for refs in references ): raise ValueError('Sacrebleu requires the same number of references for each prediction' ) _lowercase : int = [[refs[i] for refs in references] for i in range(_lowerCAmelCase )] _lowercase : int = TER( normalized=_lowerCAmelCase , no_punct=_lowerCAmelCase , asian_support=_lowerCAmelCase , case_sensitive=_lowerCAmelCase , ) _lowercase : List[Any] = sb_ter.corpus_score(_lowerCAmelCase , _lowerCAmelCase ) return {"score": output.score, "num_edits": output.num_edits, "ref_length": output.ref_length}
66
from __future__ import annotations UpperCamelCase = tuple[int, int, int] UpperCamelCase = tuple[str, str, str] # used alphabet -------------------------- # from string.ascii_uppercase UpperCamelCase = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" # -------------------------- default selection -------------------------- # rotors -------------------------- UpperCamelCase = "EGZWVONAHDCLFQMSIPJBYUKXTR" UpperCamelCase = "FOBHMDKEXQNRAULPGSJVTYICZW" UpperCamelCase = "ZJXESIUQLHAVRMDOYGTNFWPBKC" # reflector -------------------------- UpperCamelCase = { "A": "N", "N": "A", "B": "O", "O": "B", "C": "P", "P": "C", "D": "Q", "Q": "D", "E": "R", "R": "E", "F": "S", "S": "F", "G": "T", "T": "G", "H": "U", "U": "H", "I": "V", "V": "I", "J": "W", "W": "J", "K": "X", "X": "K", "L": "Y", "Y": "L", "M": "Z", "Z": "M", } # -------------------------- extra rotors -------------------------- UpperCamelCase = "RMDJXFUWGISLHVTCQNKYPBEZOA" UpperCamelCase = "SGLCPQWZHKXAREONTFBVIYJUDM" UpperCamelCase = "HVSICLTYKQUBXDWAJZOMFGPREN" UpperCamelCase = "RZWQHFMVDBKICJLNTUXAGYPSOE" UpperCamelCase = "LFKIJODBEGAMQPXVUHYSTCZRWN" UpperCamelCase = "KOAEGVDHXPQZMLFTYWJNBRCIUS" def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> tuple[RotorPositionT, RotorSelectionT, dict[str, str]]: # Checks if there are 3 unique rotors if (unique_rotsel := len(set(SCREAMING_SNAKE_CASE ) )) < 3: _lowercase : Optional[int] = F"""Please use 3 unique rotors (not {unique_rotsel})""" raise Exception(SCREAMING_SNAKE_CASE ) # Checks if rotor positions are valid _lowercase , _lowercase , _lowercase : int = rotpos if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : Dict = F"""First rotor position is not within range of 1..26 ({rotorposa}""" raise ValueError(SCREAMING_SNAKE_CASE ) if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : int = F"""Second rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(SCREAMING_SNAKE_CASE ) if not 0 < rotorposa <= len(SCREAMING_SNAKE_CASE ): _lowercase : str = F"""Third rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(SCREAMING_SNAKE_CASE ) # Validates string and returns dict _lowercase : Tuple = _plugboard(SCREAMING_SNAKE_CASE ) return rotpos, rotsel, pbdict def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> dict[str, str]: # tests the input string if it # a) is type string # b) has even length (so pairs can be made) if not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): _lowercase : Optional[int] = F"""Plugboard setting isn't type string ({type(SCREAMING_SNAKE_CASE )})""" raise TypeError(SCREAMING_SNAKE_CASE ) elif len(SCREAMING_SNAKE_CASE ) % 2 != 0: _lowercase : Optional[int] = F"""Odd number of symbols ({len(SCREAMING_SNAKE_CASE )})""" raise Exception(SCREAMING_SNAKE_CASE ) elif pbstring == "": return {} pbstring.replace(' ' , '' ) # Checks if all characters are unique _lowercase : Dict = set() for i in pbstring: if i not in abc: _lowercase : str = F"""'{i}' not in list of symbols""" raise Exception(SCREAMING_SNAKE_CASE ) elif i in tmppbl: _lowercase : int = F"""Duplicate symbol ({i})""" raise Exception(SCREAMING_SNAKE_CASE ) else: tmppbl.add(SCREAMING_SNAKE_CASE ) del tmppbl # Created the dictionary _lowercase : Optional[Any] = {} for j in range(0 , len(SCREAMING_SNAKE_CASE ) - 1 , 2 ): _lowercase : Dict = pbstring[j + 1] _lowercase : Union[str, Any] = pbstring[j] return pb def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = (rotora, rotora, rotora) , SCREAMING_SNAKE_CASE = "" , ) -> str: _lowercase : List[str] = text.upper() _lowercase , _lowercase , _lowercase : List[str] = _validator( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , plugb.upper() ) _lowercase , _lowercase , _lowercase : Optional[int] = rotor_position _lowercase , _lowercase , _lowercase : Union[str, Any] = rotor_selection rotorposa -= 1 rotorposa -= 1 rotorposa -= 1 _lowercase : Optional[int] = [] # encryption/decryption process -------------------------- for symbol in text: if symbol in abc: # 1st plugboard -------------------------- if symbol in plugboard: _lowercase : Dict = plugboard[symbol] # rotor ra -------------------------- _lowercase : Optional[Any] = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : Union[str, Any] = rotora[index % len(SCREAMING_SNAKE_CASE )] # rotor rb -------------------------- _lowercase : Tuple = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : str = rotora[index % len(SCREAMING_SNAKE_CASE )] # rotor rc -------------------------- _lowercase : List[Any] = abc.index(SCREAMING_SNAKE_CASE ) + rotorposa _lowercase : List[str] = rotora[index % len(SCREAMING_SNAKE_CASE )] # reflector -------------------------- # this is the reason you don't need another machine to decipher _lowercase : List[str] = reflector[symbol] # 2nd rotors _lowercase : List[str] = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] _lowercase : Tuple = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] _lowercase : Dict = abc[rotora.index(SCREAMING_SNAKE_CASE ) - rotorposa] # 2nd plugboard if symbol in plugboard: _lowercase : int = plugboard[symbol] # moves/resets rotor positions rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : Any = 0 rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : int = 0 rotorposa += 1 if rotorposa >= len(SCREAMING_SNAKE_CASE ): _lowercase : Any = 0 # else: # pass # Error could be also raised # raise ValueError( # 'Invalid symbol('+repr(symbol)+')') result.append(SCREAMING_SNAKE_CASE ) return "".join(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": UpperCamelCase = "This is my Python script that emulates the Enigma machine from WWII." UpperCamelCase = (1, 1, 1) UpperCamelCase = "pictures" UpperCamelCase = (rotora, rotora, rotora) UpperCamelCase = enigma(message, rotor_pos, rotor_sel, pb) print("Encrypted message:", en) print("Decrypted message:", enigma(en, rotor_pos, rotor_sel, pb))
66
1
from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging UpperCamelCase = logging.get_logger(__name__) # pylint: disable=invalid-name class lowerCAmelCase_ ( __snake_case ): def __init__( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , ): super().__init__() if hasattr(scheduler.config , 'steps_offset' ) and scheduler.config.steps_offset != 1: _lowercase : Dict = ( F"""The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`""" F""" should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure """ 'to update the config accordingly as leaving `steps_offset` might led to incorrect results' ' in future versions. If you have downloaded this checkpoint from the Hugging Face Hub,' ' it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`' ' file' ) deprecate('steps_offset!=1' , '1.0.0' , _lowerCAmelCase , standard_warn=_lowerCAmelCase ) _lowercase : List[str] = dict(scheduler.config ) _lowercase : Any = 1 _lowercase : List[Any] = FrozenDict(_lowerCAmelCase ) if hasattr(scheduler.config , 'skip_prk_steps' ) and scheduler.config.skip_prk_steps is False: _lowercase : Any = ( F"""The configuration file of this scheduler: {scheduler} has not set the configuration""" ' `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make' ' sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to' ' incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face' ' Hub, it would be very nice if you could open a Pull request for the' ' `scheduler/scheduler_config.json` file' ) deprecate('skip_prk_steps not set' , '1.0.0' , _lowerCAmelCase , standard_warn=_lowerCAmelCase ) _lowercase : str = dict(scheduler.config ) _lowercase : int = True _lowercase : Union[str, Any] = FrozenDict(_lowerCAmelCase ) if safety_checker is None: logger.warning( F"""You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure""" ' that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered' ' results in services or applications open to the public. Both the diffusers team and Hugging Face' ' strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling' ' it only for use-cases that involve analyzing network behavior or auditing its results. For more' ' information, please have a look at https://github.com/huggingface/diffusers/pull/254 .' ) self.register_modules( segmentation_model=_lowerCAmelCase , segmentation_processor=_lowerCAmelCase , vae=_lowerCAmelCase , text_encoder=_lowerCAmelCase , tokenizer=_lowerCAmelCase , unet=_lowerCAmelCase , scheduler=_lowerCAmelCase , safety_checker=_lowerCAmelCase , feature_extractor=_lowerCAmelCase , ) def __a ( self , _lowerCAmelCase = "auto" ): if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory _lowercase : str = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(_lowerCAmelCase ) def __a ( self ): self.enable_attention_slicing(_lowerCAmelCase ) def __a ( self ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) _lowercase : Tuple = torch.device('cuda' ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(_lowerCAmelCase , _lowerCAmelCase ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __a ( self ): if self.device != torch.device('meta' ) or not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(_lowerCAmelCase , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = 5_1_2 , _lowerCAmelCase = 5_1_2 , _lowerCAmelCase = 5_0 , _lowerCAmelCase = 7.5 , _lowerCAmelCase = None , _lowerCAmelCase = 1 , _lowerCAmelCase = 0.0 , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = "pil" , _lowerCAmelCase = True , _lowerCAmelCase = None , _lowerCAmelCase = 1 , **_lowerCAmelCase , ): _lowercase : str = self.segmentation_processor( text=[text] , images=[image] , padding='max_length' , return_tensors='pt' ).to(self.device ) _lowercase : Any = self.segmentation_model(**_lowerCAmelCase ) _lowercase : str = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() _lowercase : Optional[int] = self.numpy_to_pil(_lowerCAmelCase )[0].resize(image.size ) # Run inpainting pipeline with the generated mask _lowercase : str = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=_lowerCAmelCase , image=_lowerCAmelCase , mask_image=_lowerCAmelCase , height=_lowerCAmelCase , width=_lowerCAmelCase , num_inference_steps=_lowerCAmelCase , guidance_scale=_lowerCAmelCase , negative_prompt=_lowerCAmelCase , num_images_per_prompt=_lowerCAmelCase , eta=_lowerCAmelCase , generator=_lowerCAmelCase , latents=_lowerCAmelCase , output_type=_lowerCAmelCase , return_dict=_lowerCAmelCase , callback=_lowerCAmelCase , callback_steps=_lowerCAmelCase , )
66
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase = {"configuration_glpn": ["GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP", "GLPNConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = ["GLPNFeatureExtractor"] UpperCamelCase = ["GLPNImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase = [ "GLPN_PRETRAINED_MODEL_ARCHIVE_LIST", "GLPNForDepthEstimation", "GLPNLayer", "GLPNModel", "GLPNPreTrainedModel", ] if TYPE_CHECKING: from .configuration_glpn import GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP, GLPNConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_glpn import GLPNFeatureExtractor from .image_processing_glpn import GLPNImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_glpn import ( GLPN_PRETRAINED_MODEL_ARCHIVE_LIST, GLPNForDepthEstimation, GLPNLayer, GLPNModel, GLPNPreTrainedModel, ) else: import sys UpperCamelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
1
import os import pytest from attr import dataclass UpperCamelCase = "us-east-1" # defaults region @dataclass class lowerCAmelCase_ : _UpperCamelCase : str _UpperCamelCase : int = "arn:aws:iam::558105141721:role/sagemaker_execution_role" _UpperCamelCase : Dict = { "task_name": "mnli", "per_device_train_batch_size": 16, "per_device_eval_batch_size": 16, "do_train": True, "do_eval": True, "do_predict": True, "output_dir": "/opt/ml/model", "overwrite_output_dir": True, "max_steps": 500, "save_steps": 5500, } _UpperCamelCase : int = {**hyperparameters, "max_steps": 1000} @property def __a ( self ): if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def __a ( self ): return F"""{self.framework}-transfromers-test""" @property def __a ( self ): return F"""./tests/sagemaker/scripts/{self.framework}""" @property def __a ( self ): if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope='class' ) def __magic_name__ ( SCREAMING_SNAKE_CASE ) -> Any: _lowercase : List[Any] = SageMakerTestEnvironment(framework=request.cls.framework )
66
import unittest from transformers import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device if is_torch_available(): import torch from transformers import AutoModelForImageClassification if is_vision_available(): from transformers import AutoImageProcessor @require_torch @require_vision class lowerCAmelCase_ ( unittest.TestCase ): @slow def __a ( self ): _lowercase : List[Any] = AutoImageProcessor.from_pretrained('microsoft/dit-base-finetuned-rvlcdip' ) _lowercase : List[Any] = AutoModelForImageClassification.from_pretrained('microsoft/dit-base-finetuned-rvlcdip' ) model.to(_lowerCAmelCase ) from datasets import load_dataset _lowercase : Union[str, Any] = load_dataset('nielsr/rvlcdip-demo' ) _lowercase : Any = dataset['train'][0]['image'].convert('RGB' ) _lowercase : List[str] = image_processor(_lowerCAmelCase , return_tensors='pt' ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): _lowercase : Dict = model(**_lowerCAmelCase ) _lowercase : Any = outputs.logits _lowercase : str = torch.Size((1, 1_6) ) self.assertEqual(logits.shape , _lowerCAmelCase ) _lowercase : Union[str, Any] = torch.tensor( [-0.41_58, -0.40_92, -0.43_47] , device=_lowerCAmelCase , dtype=torch.float , ) self.assertTrue(torch.allclose(logits[0, :3] , _lowerCAmelCase , atol=1E-4 ) )
66
1
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, CycleDiffusionPipeline, DDIMScheduler, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowerCAmelCase_ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCamelCase : Optional[Any] = CycleDiffusionPipeline _UpperCamelCase : List[str] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { "negative_prompt", "height", "width", "negative_prompt_embeds", } _UpperCamelCase : Union[str, Any] = PipelineTesterMixin.required_optional_params - {"latents"} _UpperCamelCase : List[str] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"source_prompt"} ) _UpperCamelCase : Optional[int] = IMAGE_TO_IMAGE_IMAGE_PARAMS _UpperCamelCase : Optional[Any] = IMAGE_TO_IMAGE_IMAGE_PARAMS def __a ( self ): torch.manual_seed(0 ) _lowercase : Optional[int] = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=3_2 , ) _lowercase : int = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='scaled_linear' , num_train_timesteps=1_0_0_0 , clip_sample=_lowerCAmelCase , set_alpha_to_one=_lowerCAmelCase , ) torch.manual_seed(0 ) _lowercase : Dict = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) _lowercase : str = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) _lowercase : List[str] = CLIPTextModel(_lowerCAmelCase ) _lowercase : Dict = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) _lowercase : Any = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def __a ( self , _lowerCAmelCase , _lowerCAmelCase=0 ): _lowercase : Any = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) _lowercase : str = image / 2 + 0.5 if str(_lowerCAmelCase ).startswith('mps' ): _lowercase : List[Any] = torch.manual_seed(_lowerCAmelCase ) else: _lowercase : Optional[int] = torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) _lowercase : Any = { 'prompt': 'An astronaut riding an elephant', 'source_prompt': 'An astronaut riding a horse', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'eta': 0.1, 'strength': 0.8, 'guidance_scale': 3, 'source_guidance_scale': 1, 'output_type': 'numpy', } return inputs def __a ( self ): _lowercase : Optional[Any] = 'cpu' # ensure determinism for the device-dependent torch.Generator _lowercase : Any = self.get_dummy_components() _lowercase : Optional[Any] = CycleDiffusionPipeline(**_lowerCAmelCase ) _lowercase : Tuple = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : List[str] = self.get_dummy_inputs(_lowerCAmelCase ) _lowercase : List[Any] = pipe(**_lowerCAmelCase ) _lowercase : Dict = output.images _lowercase : Tuple = images[0, -3:, -3:, -1] assert images.shape == (1, 3_2, 3_2, 3) _lowercase : Any = np.array([0.44_59, 0.49_43, 0.45_44, 0.66_43, 0.54_74, 0.43_27, 0.57_01, 0.59_59, 0.51_79] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @unittest.skipIf(torch_device != 'cuda' , 'This test requires a GPU' ) def __a ( self ): _lowercase : Optional[int] = self.get_dummy_components() for name, module in components.items(): if hasattr(_lowerCAmelCase , 'half' ): _lowercase : Any = module.half() _lowercase : str = CycleDiffusionPipeline(**_lowerCAmelCase ) _lowercase : str = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) _lowercase : Any = self.get_dummy_inputs(_lowerCAmelCase ) _lowercase : str = pipe(**_lowerCAmelCase ) _lowercase : List[Any] = output.images _lowercase : str = images[0, -3:, -3:, -1] assert images.shape == (1, 3_2, 3_2, 3) _lowercase : Optional[Any] = np.array([0.35_06, 0.45_43, 0.4_46, 0.45_75, 0.51_95, 0.41_55, 0.52_73, 0.5_18, 0.41_16] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @skip_mps def __a ( self ): return super().test_save_load_local() @unittest.skip('non-deterministic pipeline' ) def __a ( self ): return super().test_inference_batch_single_identical() @skip_mps def __a ( self ): return super().test_dict_tuple_outputs_equivalent() @skip_mps def __a ( self ): return super().test_save_load_optional_components() @skip_mps def __a ( self ): return super().test_attention_slicing_forward_pass() @slow @require_torch_gpu class lowerCAmelCase_ ( unittest.TestCase ): def __a ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self ): _lowercase : int = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/cycle-diffusion/black_colored_car.png' ) _lowercase : List[str] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car_fp16.npy' ) _lowercase : List[str] = init_image.resize((5_1_2, 5_1_2) ) _lowercase : Any = 'CompVis/stable-diffusion-v1-4' _lowercase : Any = DDIMScheduler.from_pretrained(_lowerCAmelCase , subfolder='scheduler' ) _lowercase : Tuple = CycleDiffusionPipeline.from_pretrained( _lowerCAmelCase , scheduler=_lowerCAmelCase , safety_checker=_lowerCAmelCase , torch_dtype=torch.floataa , revision='fp16' ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) pipe.enable_attention_slicing() _lowercase : Optional[Any] = 'A black colored car' _lowercase : Union[str, Any] = 'A blue colored car' _lowercase : List[str] = torch.manual_seed(0 ) _lowercase : Union[str, Any] = pipe( prompt=_lowerCAmelCase , source_prompt=_lowerCAmelCase , image=_lowerCAmelCase , num_inference_steps=1_0_0 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=_lowerCAmelCase , output_type='np' , ) _lowercase : List[str] = output.images # the values aren't exactly equal, but the images look the same visually assert np.abs(image - expected_image ).max() < 5E-1 def __a ( self ): _lowercase : int = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/cycle-diffusion/black_colored_car.png' ) _lowercase : str = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car.npy' ) _lowercase : List[str] = init_image.resize((5_1_2, 5_1_2) ) _lowercase : Optional[int] = 'CompVis/stable-diffusion-v1-4' _lowercase : Optional[Any] = DDIMScheduler.from_pretrained(_lowerCAmelCase , subfolder='scheduler' ) _lowercase : List[str] = CycleDiffusionPipeline.from_pretrained(_lowerCAmelCase , scheduler=_lowerCAmelCase , safety_checker=_lowerCAmelCase ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) pipe.enable_attention_slicing() _lowercase : Tuple = 'A black colored car' _lowercase : Optional[Any] = 'A blue colored car' _lowercase : Optional[Any] = torch.manual_seed(0 ) _lowercase : Union[str, Any] = pipe( prompt=_lowerCAmelCase , source_prompt=_lowerCAmelCase , image=_lowerCAmelCase , num_inference_steps=1_0_0 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=_lowerCAmelCase , output_type='np' , ) _lowercase : List[str] = output.images assert np.abs(image - expected_image ).max() < 2E-2
66
from PIL import Image def __magic_name__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Image: def brightness(SCREAMING_SNAKE_CASE ) -> float: return 128 + level + (c - 128) if not -255.0 <= level <= 255.0: raise ValueError('level must be between -255.0 (black) and 255.0 (white)' ) return img.point(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": # Load image with Image.open("image_data/lena.jpg") as img: # Change brightness to 100 UpperCamelCase = change_brightness(img, 100) brigt_img.save("image_data/lena_brightness.png", format="png")
66
1