code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_donut import DonutImageProcessor lowercase__ = logging.get_logger(__name__) class lowerCAmelCase__ ( lowercase ): '''simple docstring''' def __init__( self , *lowercase , **lowercase ): warnings.warn( 'The class DonutFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use DonutImageProcessor instead.' , lowercase , ) super().__init__(*lowercase , **lowercase )
96
"""simple docstring""" lowercase__ = { """meter""": """m""", """kilometer""": """km""", """megametre""": """Mm""", """gigametre""": """Gm""", """terametre""": """Tm""", """petametre""": """Pm""", """exametre""": """Em""", """zettametre""": """Zm""", """yottametre""": """Ym""", } # Exponent of the factor(meter) lowercase__ = { """m""": 0, """km""": 3, """Mm""": 6, """Gm""": 9, """Tm""": 12, """Pm""": 15, """Em""": 18, """Zm""": 21, """Ym""": 24, } def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : List[Any] = from_type.lower().strip('s' ) _lowerCamelCase : List[Any] = to_type.lower().strip('s' ) _lowerCamelCase : Optional[int] = UNIT_SYMBOL.get(lowercase__ , lowercase__ ) _lowerCamelCase : Any = UNIT_SYMBOL.get(lowercase__ , lowercase__ ) if from_sanitized not in METRIC_CONVERSION: _lowerCamelCase : Tuple = ( f'''Invalid \'from_type\' value: {from_type!r}.\n''' f'''Conversion abbreviations are: {', '.join(lowercase__ )}''' ) raise ValueError(lowercase__ ) if to_sanitized not in METRIC_CONVERSION: _lowerCamelCase : Any = ( f'''Invalid \'to_type\' value: {to_type!r}.\n''' f'''Conversion abbreviations are: {', '.join(lowercase__ )}''' ) raise ValueError(lowercase__ ) _lowerCamelCase : List[Any] = METRIC_CONVERSION[from_sanitized] _lowerCamelCase : int = METRIC_CONVERSION[to_sanitized] _lowerCamelCase : List[str] = 1 if from_exponent > to_exponent: _lowerCamelCase : List[str] = from_exponent - to_exponent else: _lowerCamelCase : List[Any] = -(to_exponent - from_exponent) return value * pow(10 , lowercase__ ) if __name__ == "__main__": from doctest import testmod testmod()
96
1
"""simple docstring""" import argparse import requests import torch from PIL import Image from transformers import SwinConfig, SwinForMaskedImageModeling, ViTImageProcessor def _snake_case ( lowercase__ ): _lowerCamelCase : Optional[int] = SwinConfig(image_size=192 ) if "base" in model_name: _lowerCamelCase : Optional[int] = 6 _lowerCamelCase : int = 128 _lowerCamelCase : List[Any] = (2, 2, 18, 2) _lowerCamelCase : Optional[int] = (4, 8, 16, 32) elif "large" in model_name: _lowerCamelCase : int = 12 _lowerCamelCase : str = 192 _lowerCamelCase : Tuple = (2, 2, 18, 2) _lowerCamelCase : Optional[int] = (6, 12, 24, 48) else: raise ValueError('Model not supported, only supports base and large variants' ) _lowerCamelCase : List[str] = window_size _lowerCamelCase : int = embed_dim _lowerCamelCase : Optional[Any] = depths _lowerCamelCase : Any = num_heads return config def _snake_case ( lowercase__ ): if "encoder.mask_token" in name: _lowerCamelCase : Dict = name.replace('encoder.mask_token' , 'embeddings.mask_token' ) if "encoder.patch_embed.proj" in name: _lowerCamelCase : Any = name.replace('encoder.patch_embed.proj' , 'embeddings.patch_embeddings.projection' ) if "encoder.patch_embed.norm" in name: _lowerCamelCase : List[str] = name.replace('encoder.patch_embed.norm' , 'embeddings.norm' ) if "attn.proj" in name: _lowerCamelCase : Tuple = name.replace('attn.proj' , 'attention.output.dense' ) if "attn" in name: _lowerCamelCase : str = name.replace('attn' , 'attention.self' ) if "norm1" in name: _lowerCamelCase : Dict = name.replace('norm1' , 'layernorm_before' ) if "norm2" in name: _lowerCamelCase : Union[str, Any] = name.replace('norm2' , 'layernorm_after' ) if "mlp.fc1" in name: _lowerCamelCase : Optional[Any] = name.replace('mlp.fc1' , 'intermediate.dense' ) if "mlp.fc2" in name: _lowerCamelCase : int = name.replace('mlp.fc2' , 'output.dense' ) if name == "encoder.norm.weight": _lowerCamelCase : int = 'layernorm.weight' if name == "encoder.norm.bias": _lowerCamelCase : List[str] = 'layernorm.bias' if "decoder" in name: pass else: _lowerCamelCase : Union[str, Any] = 'swin.' + name return name def _snake_case ( lowercase__ , lowercase__ ): for key in orig_state_dict.copy().keys(): _lowerCamelCase : Any = orig_state_dict.pop(lowercase__ ) if "attn_mask" in key: pass elif "qkv" in key: _lowerCamelCase : List[Any] = key.split('.' ) _lowerCamelCase : Dict = int(key_split[2] ) _lowerCamelCase : List[Any] = int(key_split[4] ) _lowerCamelCase : Optional[Any] = model.swin.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: _lowerCamelCase : Optional[Any] = val[:dim, :] _lowerCamelCase : Any = val[ dim : dim * 2, : ] _lowerCamelCase : Optional[Any] = val[-dim:, :] else: _lowerCamelCase : str = val[ :dim ] _lowerCamelCase : Optional[int] = val[ dim : dim * 2 ] _lowerCamelCase : List[Any] = val[ -dim: ] else: _lowerCamelCase : Optional[int] = val return orig_state_dict def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : Optional[Any] = torch.load(lowercase__ , map_location='cpu' )['model'] _lowerCamelCase : Optional[Any] = get_swin_config(lowercase__ ) _lowerCamelCase : Optional[int] = SwinForMaskedImageModeling(lowercase__ ) model.eval() _lowerCamelCase : List[Any] = convert_state_dict(lowercase__ , lowercase__ ) model.load_state_dict(lowercase__ ) _lowerCamelCase : int = 'http://images.cocodataset.org/val2017/000000039769.jpg' _lowerCamelCase : int = ViTImageProcessor(size={'height': 192, 'width': 192} ) _lowerCamelCase : Union[str, Any] = Image.open(requests.get(lowercase__ , stream=lowercase__ ).raw ) _lowerCamelCase : Union[str, Any] = image_processor(images=lowercase__ , return_tensors='pt' ) with torch.no_grad(): _lowerCamelCase : List[Any] = model(**lowercase__ ).logits print(outputs.keys() ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(f'''Saving model {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowercase__ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(lowercase__ ) if push_to_hub: print(f'''Pushing model and image processor for {model_name} to hub''' ) model.push_to_hub(f'''microsoft/{model_name}''' ) image_processor.push_to_hub(f'''microsoft/{model_name}''' ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""swin-base-simmim-window6-192""", type=str, choices=["""swin-base-simmim-window6-192""", """swin-large-simmim-window12-192"""], help="""Name of the Swin SimMIM model you'd like to convert.""", ) parser.add_argument( """--checkpoint_path""", default="""/Users/nielsrogge/Documents/SwinSimMIM/simmim_pretrain__swin_base__img192_window6__100ep.pth""", type=str, help="""Path to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) lowercase__ = parser.parse_args() convert_swin_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
96
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ = {"""configuration_vit_msn""": ["""VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ViTMSNConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ """VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST""", """ViTMSNModel""", """ViTMSNForImageClassification""", """ViTMSNPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_vit_msn import VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMSNConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_msn import ( VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMSNForImageClassification, ViTMSNModel, ViTMSNPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
96
1
"""simple docstring""" import argparse import shutil import time from json import JSONDecodeError from logging import getLogger from pathlib import Path from typing import Dict, List import torch from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from utils import ( SeqaSeqDataset, calculate_bleu, calculate_rouge, chunks, lmap, load_json, parse_numeric_n_bool_cl_kwargs, save_json, use_task_specific_params, write_txt_file, ) lowercase__ = getLogger(__name__) def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ = 8 , lowercase__ = 1024 , lowercase__="val" , lowercase__=None , lowercase__=False , lowercase__="summarization" , lowercase__=None , lowercase__=1 , lowercase__ = None , lowercase__="" , **lowercase__ , ): _lowerCamelCase : List[str] = str(lowercase__ ) assert local_rank is not None torch.distributed.init_process_group(backend='nccl' , rank=lowercase__ ) _lowerCamelCase : str = Path(lowercase__ ) _lowerCamelCase : int = save_dir.joinpath(f'''rank_{local_rank}_output.json''' ) torch.cuda.set_device(lowercase__ ) _lowerCamelCase : Any = AutoModelForSeqaSeqLM.from_pretrained(lowercase__ ).cuda() if fpaa: _lowerCamelCase : Optional[Any] = model.half() # determine if we need to increase num_beams use_task_specific_params(lowercase__ , lowercase__ ) # update config with task specific params _lowerCamelCase : Any = generate_kwargs.pop('num_beams' , model.config.num_beams ) # AttributeError risk? if num_return_sequences > num_beams: _lowerCamelCase : Dict = num_return_sequences _lowerCamelCase : List[Any] = AutoTokenizer.from_pretrained(lowercase__ ) logger.info(f'''Inferred tokenizer type: {tokenizer.__class__}''' ) # if this is wrong, check config.model_type. if max_source_length is None: _lowerCamelCase : Tuple = tokenizer.model_max_length if prefix is None: _lowerCamelCase : Optional[Any] = prefix or getattr(model.config , 'prefix' , '' ) or '' _lowerCamelCase : int = SeqaSeqDataset( lowercase__ , lowercase__ , lowercase__ , max_target_length=1024 , type_path=lowercase__ , n_obs=lowercase__ , prefix=lowercase__ , **lowercase__ , ) # I set shuffle=True for a more accurate progress bar. # If all the longest samples are first, the prog bar estimate is too high at the beginning. _lowerCamelCase : List[Any] = ds.make_sortish_sampler(lowercase__ , distributed=lowercase__ , add_extra_examples=lowercase__ , shuffle=lowercase__ ) _lowerCamelCase : List[Any] = DataLoader(lowercase__ , sampler=lowercase__ , batch_size=lowercase__ , collate_fn=ds.collate_fn ) _lowerCamelCase : Optional[Any] = [] for batch in tqdm(lowercase__ ): _lowerCamelCase : Tuple = model.generate( input_ids=batch['input_ids'].to(model.device ) , attention_mask=batch['attention_mask'].to(model.device ) , num_return_sequences=lowercase__ , num_beams=lowercase__ , **lowercase__ , ) _lowerCamelCase : Tuple = tokenizer.batch_decode(lowercase__ , skip_special_tokens=lowercase__ , clean_up_tokenization_spaces=lowercase__ ) _lowerCamelCase : str = batch['ids'] if num_return_sequences > 1: _lowerCamelCase : Any = chunks(lowercase__ , lowercase__ ) # batch size chunks, each of size num_return_seq for i, pred in enumerate(lowercase__ ): results.append({'pred': pred, 'id': ids[i].item()} ) save_json(lowercase__ , lowercase__ ) return results, sampler.num_replicas def _snake_case ( ): _lowerCamelCase : str = argparse.ArgumentParser( epilog='Unspecified args like --num_beams=2 --decoder_start_token_id=4 are passed to model.generate' ) parser.add_argument('--data_dir' , type=lowercase__ , help='like cnn_dm/test.source' ) parser.add_argument( '--model_name' , type=lowercase__ , help='like facebook/bart-large-cnn,t5-base, etc.' , default='sshleifer/distilbart-xsum-12-3' , ) parser.add_argument('--save_dir' , type=lowercase__ , help='where to save' , default='tmp_gen' ) parser.add_argument('--max_source_length' , type=lowercase__ , default=lowercase__ ) parser.add_argument( '--type_path' , type=lowercase__ , default='test' , help='which subset to evaluate typically train/val/test' ) parser.add_argument('--task' , type=lowercase__ , default='summarization' , help='used for task_specific_params + metrics' ) parser.add_argument('--bs' , type=lowercase__ , default=8 , required=lowercase__ , help='batch size' ) parser.add_argument( '--local_rank' , type=lowercase__ , default=-1 , required=lowercase__ , help='should be passed by distributed.launch' ) parser.add_argument( '--n_obs' , type=lowercase__ , default=lowercase__ , required=lowercase__ , help='How many observations. Defaults to all.' ) parser.add_argument( '--num_return_sequences' , type=lowercase__ , default=1 , required=lowercase__ , help='How many sequences to return' ) parser.add_argument( '--sync_timeout' , type=lowercase__ , default=600 , required=lowercase__ , help='How long should master process wait for other processes to finish.' , ) parser.add_argument('--src_lang' , type=lowercase__ , default=lowercase__ , required=lowercase__ ) parser.add_argument('--tgt_lang' , type=lowercase__ , default=lowercase__ , required=lowercase__ ) parser.add_argument( '--prefix' , type=lowercase__ , required=lowercase__ , default=lowercase__ , help='will be added to the begininng of src examples' ) parser.add_argument('--fp16' , action='store_true' ) parser.add_argument('--debug' , action='store_true' ) _lowerCamelCase : str = time.time() _lowerCamelCase, _lowerCamelCase : List[Any] = parser.parse_known_args() _lowerCamelCase : Any = parse_numeric_n_bool_cl_kwargs(lowercase__ ) if generate_kwargs and args.local_rank <= 0: print(f'''parsed the following generate kwargs: {generate_kwargs}''' ) _lowerCamelCase : List[Any] = Path(args.save_dir + '_tmp' ) Path(lowercase__ ).mkdir(exist_ok=lowercase__ ) # this handles locking. _lowerCamelCase : Any = list(json_save_dir.glob('rank_*.json' ) ) if intermediate_files: raise ValueError(f'''Found files at {json_save_dir} please move or remove them.''' ) # In theory, a node could finish and save before another node hits this. If this happens, we can address later. _lowerCamelCase : Optional[Any] = {} if args.src_lang is not None: _lowerCamelCase : Tuple = args.src_lang if args.tgt_lang is not None: _lowerCamelCase : Tuple = args.tgt_lang Path(args.save_dir ).mkdir(exist_ok=lowercase__ ) _lowerCamelCase, _lowerCamelCase : Optional[Any] = eval_data_dir( args.data_dir , lowercase__ , args.model_name , type_path=args.type_path , bs=args.bs , fpaa=args.fpaa , task=args.task , local_rank=args.local_rank , n_obs=args.n_obs , max_source_length=args.max_source_length , num_return_sequences=args.num_return_sequences , prefix=args.prefix , dataset_kwargs=lowercase__ , **lowercase__ , ) if args.local_rank <= 0: _lowerCamelCase : Tuple = Path(args.save_dir ) save_dir.mkdir(exist_ok=lowercase__ ) _lowerCamelCase : Union[str, Any] = gather_results_from_each_node(lowercase__ , lowercase__ , args.sync_timeout ) _lowerCamelCase : str = combine_partial_results(lowercase__ ) if args.num_return_sequences > 1: _lowerCamelCase : str = save_dir.joinpath('pseudolabel_results.json' ) print(f'''Saving aggregated results at {save_path}, intermediate in {json_save_dir}/''' ) save_json(lowercase__ , lowercase__ ) return _lowerCamelCase : Any = Path(args.data_dir ).joinpath(args.type_path + '.target' ) with open(lowercase__ ) as f: _lowerCamelCase : int = [x.rstrip() for x in f.readlines()][: len(lowercase__ )] # Calculate metrics, save metrics, and save _generations.txt _lowerCamelCase : str = 'translation' in args.task _lowerCamelCase : Tuple = calculate_bleu if calc_bleu else calculate_rouge _lowerCamelCase : List[str] = 'bleu' if calc_bleu else 'rouge' _lowerCamelCase : Dict = score_fn(lowercase__ , lowercase__ ) _lowerCamelCase : Any = len(lowercase__ ) _lowerCamelCase : Dict = time.time() - start_time _lowerCamelCase : Optional[int] = round(runtime / metrics['n_obs'] , 4 ) _lowerCamelCase : Optional[Any] = num_replicas # TODO(@stas00): add whatever metadata to metrics _lowerCamelCase : Union[str, Any] = save_dir.joinpath(f'''{args.type_path}_{metric_name}.json''' ) save_json(lowercase__ , lowercase__ , indent=lowercase__ ) print(lowercase__ ) write_txt_file(lowercase__ , save_dir.joinpath(f'''{args.type_path}_generations.txt''' ) ) if args.debug: write_txt_file(lowercase__ , save_dir.joinpath(f'''{args.type_path}.target''' ) ) else: shutil.rmtree(lowercase__ ) def _snake_case ( lowercase__ ): _lowerCamelCase : str = [] for partial_result in partial_results: records.extend(lowercase__ ) _lowerCamelCase : int = sorted(lowercase__ , key=lambda lowercase__ : x["id"] ) _lowerCamelCase : List[Any] = [x['pred'] for x in records] return preds def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): # WAIT FOR lots of .json files _lowerCamelCase : str = time.time() logger.info('waiting for all nodes to finish' ) _lowerCamelCase : Dict = None while (time.time() - start_wait) < timeout: _lowerCamelCase : Union[str, Any] = list(save_dir.glob('rank_*.json' ) ) if len(lowercase__ ) < num_replicas: continue try: # make sure all json files are fully saved _lowerCamelCase : str = lmap(lowercase__ , lowercase__ ) return json_data except JSONDecodeError: continue else: raise TimeoutError('Rank 0 gave up on waiting for other processes' ) # Unreachable if __name__ == "__main__": # Usage for MT: run_generate()
96
"""simple docstring""" def _snake_case ( lowercase__ , lowercase__ ): # "extended trapezoidal rule" # int(f) = dx/2 * (f1 + 2f2 + ... + fn) _lowerCamelCase : List[Any] = (boundary[1] - boundary[0]) / steps _lowerCamelCase : Tuple = boundary[0] _lowerCamelCase : Dict = boundary[1] _lowerCamelCase : List[Any] = make_points(lowercase__ , lowercase__ , lowercase__ ) _lowerCamelCase : List[Any] = 0.0 y += (h / 2.0) * f(lowercase__ ) for i in x_i: # print(i) y += h * f(lowercase__ ) y += (h / 2.0) * f(lowercase__ ) return y def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : str = a + h while x < (b - h): yield x _lowerCamelCase : int = x + h def _snake_case ( lowercase__ ): # enter your function here _lowerCamelCase : Optional[Any] = (x - 0) * (x - 0) return y def _snake_case ( ): _lowerCamelCase : int = 0.0 # Lower bound of integration _lowerCamelCase : Optional[int] = 1.0 # Upper bound of integration _lowerCamelCase : List[str] = 1_0.0 # define number of steps or resolution _lowerCamelCase : List[Any] = [a, b] # define boundary of integration _lowerCamelCase : Optional[Any] = method_a(lowercase__ , lowercase__ ) print(f'''y = {y}''' ) if __name__ == "__main__": main()
96
1
"""simple docstring""" from collections.abc import Sequence from queue import Queue class lowerCAmelCase__ : '''simple docstring''' def __init__( self , lowercase , lowercase , lowercase , lowercase=None , lowercase=None ): _lowerCamelCase : Any = start _lowerCamelCase : Optional[Any] = end _lowerCamelCase : Optional[int] = val _lowerCamelCase : List[str] = (start + end) // 2 _lowerCamelCase : Any = left _lowerCamelCase : Tuple = right def __repr__( self ): return F'''SegmentTreeNode(start={self.start}, end={self.end}, val={self.val})''' class lowerCAmelCase__ : '''simple docstring''' def __init__( self , lowercase , lowercase ): _lowerCamelCase : List[Any] = collection _lowerCamelCase : Tuple = function if self.collection: _lowerCamelCase : str = self._build_tree(0 , len(lowercase ) - 1 ) def A_ ( self , lowercase , lowercase ): self._update_tree(self.root , lowercase , lowercase ) def A_ ( self , lowercase , lowercase ): return self._query_range(self.root , lowercase , lowercase ) def A_ ( self , lowercase , lowercase ): if start == end: return SegmentTreeNode(lowercase , lowercase , self.collection[start] ) _lowerCamelCase : int = (start + end) // 2 _lowerCamelCase : Any = self._build_tree(lowercase , lowercase ) _lowerCamelCase : Union[str, Any] = self._build_tree(mid + 1 , lowercase ) return SegmentTreeNode(lowercase , lowercase , self.fn(left.val , right.val ) , lowercase , lowercase ) def A_ ( self , lowercase , lowercase , lowercase ): if node.start == i and node.end == i: _lowerCamelCase : Any = val return if i <= node.mid: self._update_tree(node.left , lowercase , lowercase ) else: self._update_tree(node.right , lowercase , lowercase ) _lowerCamelCase : List[str] = self.fn(node.left.val , node.right.val ) def A_ ( self , lowercase , lowercase , lowercase ): if node.start == i and node.end == j: return node.val if i <= node.mid: if j <= node.mid: # range in left child tree return self._query_range(node.left , lowercase , lowercase ) else: # range in left child tree and right child tree return self.fn( self._query_range(node.left , lowercase , node.mid ) , self._query_range(node.right , node.mid + 1 , lowercase ) , ) else: # range in right child tree return self._query_range(node.right , lowercase , lowercase ) def A_ ( self ): if self.root is not None: _lowerCamelCase : List[str] = Queue() queue.put(self.root ) while not queue.empty(): _lowerCamelCase : List[str] = queue.get() yield node if node.left is not None: queue.put(node.left ) if node.right is not None: queue.put(node.right ) if __name__ == "__main__": import operator for fn in [operator.add, max, min]: print("""*""" * 50) lowercase__ = SegmentTree([2, 1, 5, 3, 4], fn) for node in arr.traverse(): print(node) print() arr.update(1, 5) for node in arr.traverse(): print(node) print() print(arr.query_range(3, 4)) # 7 print(arr.query_range(2, 2)) # 5 print(arr.query_range(1, 3)) # 13 print()
96
"""simple docstring""" import math def _snake_case ( lowercase__ ): return math.sqrt(lowercase__ ) * math.sqrt(lowercase__ ) == num def _snake_case ( lowercase__ ): _lowerCamelCase : Optional[int] = 0 _lowerCamelCase : List[Any] = n while left <= right: _lowerCamelCase : str = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: _lowerCamelCase : str = mid - 1 else: _lowerCamelCase : Optional[int] = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" from __future__ import annotations import typing from collections.abc import Iterable import numpy as np lowercase__ = typing.Union[Iterable[float], Iterable[int], np.ndarray] # noqa: UP007 lowercase__ = typing.Union[np.floataa, int, float] # noqa: UP007 def _snake_case ( lowercase__ , lowercase__ ): return np.sqrt(np.sum((np.asarray(lowercase__ ) - np.asarray(lowercase__ )) ** 2 ) ) def _snake_case ( lowercase__ , lowercase__ ): return sum((va - va) ** 2 for va, va in zip(lowercase__ , lowercase__ ) ) ** (1 / 2) if __name__ == "__main__": def _snake_case ( ): from timeit import timeit print('Without Numpy' ) print( timeit( 'euclidean_distance_no_np([1, 2, 3], [4, 5, 6])' , number=10000 , globals=globals() , ) ) print('With Numpy' ) print( timeit( 'euclidean_distance([1, 2, 3], [4, 5, 6])' , number=10000 , globals=globals() , ) ) benchmark()
96
"""simple docstring""" import functools from typing import Any def _snake_case ( lowercase__ , lowercase__ ): # Validation if not isinstance(lowercase__ , lowercase__ ) or len(lowercase__ ) == 0: raise ValueError('the string should be not empty string' ) if not isinstance(lowercase__ , lowercase__ ) or not all( isinstance(lowercase__ , lowercase__ ) and len(lowercase__ ) > 0 for item in words ): raise ValueError('the words should be a list of non-empty strings' ) # Build trie _lowerCamelCase : dict[str, Any] = {} _lowerCamelCase : List[Any] = 'WORD_KEEPER' for word in words: _lowerCamelCase : Dict = trie for c in word: if c not in trie_node: _lowerCamelCase : Any = {} _lowerCamelCase : str = trie_node[c] _lowerCamelCase : Optional[Any] = True _lowerCamelCase : Dict = len(lowercase__ ) # Dynamic programming method @functools.cache def is_breakable(lowercase__ ) -> bool: if index == len_string: return True _lowerCamelCase : List[Any] = trie for i in range(lowercase__ , lowercase__ ): _lowerCamelCase : Any = trie_node.get(string[i] , lowercase__ ) if trie_node is None: return False if trie_node.get(lowercase__ , lowercase__ ) and is_breakable(i + 1 ): return True return False return is_breakable(0 ) if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" import json import os import torch from diffusers import UNetaDModel os.makedirs("""hub/hopper-medium-v2/unet/hor32""", exist_ok=True) os.makedirs("""hub/hopper-medium-v2/unet/hor128""", exist_ok=True) os.makedirs("""hub/hopper-medium-v2/value_function""", exist_ok=True) def _snake_case ( lowercase__ ): if hor == 128: _lowerCamelCase : Dict = ('DownResnetBlock1D', 'DownResnetBlock1D', 'DownResnetBlock1D') _lowerCamelCase : List[Any] = (32, 128, 256) _lowerCamelCase : List[str] = ('UpResnetBlock1D', 'UpResnetBlock1D') elif hor == 32: _lowerCamelCase : Optional[Any] = ('DownResnetBlock1D', 'DownResnetBlock1D', 'DownResnetBlock1D', 'DownResnetBlock1D') _lowerCamelCase : List[Any] = (32, 64, 128, 256) _lowerCamelCase : Any = ('UpResnetBlock1D', 'UpResnetBlock1D', 'UpResnetBlock1D') _lowerCamelCase : Dict = torch.load(f'''/Users/bglickenhaus/Documents/diffuser/temporal_unet-hopper-mediumv2-hor{hor}.torch''' ) _lowerCamelCase : Dict = model.state_dict() _lowerCamelCase : List[str] = { 'down_block_types': down_block_types, 'block_out_channels': block_out_channels, 'up_block_types': up_block_types, 'layers_per_block': 1, 'use_timestep_embedding': True, 'out_block_type': 'OutConv1DBlock', 'norm_num_groups': 8, 'downsample_each_block': False, 'in_channels': 14, 'out_channels': 14, 'extra_in_channels': 0, 'time_embedding_type': 'positional', 'flip_sin_to_cos': False, 'freq_shift': 1, 'sample_size': 65536, 'mid_block_type': 'MidResTemporalBlock1D', 'act_fn': 'mish', } _lowerCamelCase : List[str] = UNetaDModel(**lowercase__ ) print(f'''length of state dict: {len(state_dict.keys() )}''' ) print(f'''length of value function dict: {len(hf_value_function.state_dict().keys() )}''' ) _lowerCamelCase : List[Any] = dict(zip(model.state_dict().keys() , hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): _lowerCamelCase : List[Any] = state_dict.pop(lowercase__ ) hf_value_function.load_state_dict(lowercase__ ) torch.save(hf_value_function.state_dict() , f'''hub/hopper-medium-v2/unet/hor{hor}/diffusion_pytorch_model.bin''' ) with open(f'''hub/hopper-medium-v2/unet/hor{hor}/config.json''' , 'w' ) as f: json.dump(lowercase__ , lowercase__ ) def _snake_case ( ): _lowerCamelCase : Tuple = { 'in_channels': 14, 'down_block_types': ('DownResnetBlock1D', 'DownResnetBlock1D', 'DownResnetBlock1D', 'DownResnetBlock1D'), 'up_block_types': (), 'out_block_type': 'ValueFunction', 'mid_block_type': 'ValueFunctionMidBlock1D', 'block_out_channels': (32, 64, 128, 256), 'layers_per_block': 1, 'downsample_each_block': True, 'sample_size': 65536, 'out_channels': 14, 'extra_in_channels': 0, 'time_embedding_type': 'positional', 'use_timestep_embedding': True, 'flip_sin_to_cos': False, 'freq_shift': 1, 'norm_num_groups': 8, 'act_fn': 'mish', } _lowerCamelCase : Optional[int] = torch.load('/Users/bglickenhaus/Documents/diffuser/value_function-hopper-mediumv2-hor32.torch' ) _lowerCamelCase : Union[str, Any] = model _lowerCamelCase : Union[str, Any] = UNetaDModel(**lowercase__ ) print(f'''length of state dict: {len(state_dict.keys() )}''' ) print(f'''length of value function dict: {len(hf_value_function.state_dict().keys() )}''' ) _lowerCamelCase : Tuple = dict(zip(state_dict.keys() , hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): _lowerCamelCase : Optional[int] = state_dict.pop(lowercase__ ) hf_value_function.load_state_dict(lowercase__ ) torch.save(hf_value_function.state_dict() , 'hub/hopper-medium-v2/value_function/diffusion_pytorch_model.bin' ) with open('hub/hopper-medium-v2/value_function/config.json' , 'w' ) as f: json.dump(lowercase__ , lowercase__ ) if __name__ == "__main__": unet(32) # unet(128) value_function()
96
"""simple docstring""" def _snake_case ( lowercase__ ): if not isinstance(lowercase__ , lowercase__ ): raise ValueError('Input series is not valid, valid series - [2, 4, 6]' ) if len(lowercase__ ) == 0: raise ValueError('Input list must be a non empty list' ) if len(lowercase__ ) == 1: return True _lowerCamelCase : List[Any] = series[1] - series[0] for index in range(len(lowercase__ ) - 1 ): if series[index + 1] - series[index] != common_diff: return False return True def _snake_case ( lowercase__ ): if not isinstance(lowercase__ , lowercase__ ): raise ValueError('Input series is not valid, valid series - [2, 4, 6]' ) if len(lowercase__ ) == 0: raise ValueError('Input list must be a non empty list' ) _lowerCamelCase : Optional[int] = 0 for val in series: answer += val return answer / len(lowercase__ ) if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" import warnings from ...utils import is_sklearn_available, requires_backends if is_sklearn_available(): from scipy.stats import pearsonr, spearmanr from sklearn.metrics import fa_score, matthews_corrcoef lowercase__ = ( """This metric will be removed from the library soon, metrics should be handled with the 🤗 Evaluate """ """library. You can have a look at this example script for pointers: """ """https://github.com/huggingface/transformers/blob/main/examples/pytorch/text-classification/run_glue.py""" ) def _snake_case ( lowercase__ , lowercase__ ): warnings.warn(lowercase__ , lowercase__ ) requires_backends(lowercase__ , 'sklearn' ) return (preds == labels).mean() def _snake_case ( lowercase__ , lowercase__ ): warnings.warn(lowercase__ , lowercase__ ) requires_backends(lowercase__ , 'sklearn' ) _lowerCamelCase : Optional[Any] = simple_accuracy(lowercase__ , lowercase__ ) _lowerCamelCase : str = fa_score(y_true=lowercase__ , y_pred=lowercase__ ) return { "acc": acc, "f1": fa, "acc_and_f1": (acc + fa) / 2, } def _snake_case ( lowercase__ , lowercase__ ): warnings.warn(lowercase__ , lowercase__ ) requires_backends(lowercase__ , 'sklearn' ) _lowerCamelCase : Optional[int] = pearsonr(lowercase__ , lowercase__ )[0] _lowerCamelCase : int = spearmanr(lowercase__ , lowercase__ )[0] return { "pearson": pearson_corr, "spearmanr": spearman_corr, "corr": (pearson_corr + spearman_corr) / 2, } def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): warnings.warn(lowercase__ , lowercase__ ) requires_backends(lowercase__ , 'sklearn' ) assert len(lowercase__ ) == len(lowercase__ ), f'''Predictions and labels have mismatched lengths {len(lowercase__ )} and {len(lowercase__ )}''' if task_name == "cola": return {"mcc": matthews_corrcoef(lowercase__ , lowercase__ )} elif task_name == "sst-2": return {"acc": simple_accuracy(lowercase__ , lowercase__ )} elif task_name == "mrpc": return acc_and_fa(lowercase__ , lowercase__ ) elif task_name == "sts-b": return pearson_and_spearman(lowercase__ , lowercase__ ) elif task_name == "qqp": return acc_and_fa(lowercase__ , lowercase__ ) elif task_name == "mnli": return {"mnli/acc": simple_accuracy(lowercase__ , lowercase__ )} elif task_name == "mnli-mm": return {"mnli-mm/acc": simple_accuracy(lowercase__ , lowercase__ )} elif task_name == "qnli": return {"acc": simple_accuracy(lowercase__ , lowercase__ )} elif task_name == "rte": return {"acc": simple_accuracy(lowercase__ , lowercase__ )} elif task_name == "wnli": return {"acc": simple_accuracy(lowercase__ , lowercase__ )} elif task_name == "hans": return {"acc": simple_accuracy(lowercase__ , lowercase__ )} else: raise KeyError(lowercase__ ) def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): warnings.warn(lowercase__ , lowercase__ ) requires_backends(lowercase__ , 'sklearn' ) if len(lowercase__ ) != len(lowercase__ ): raise ValueError(f'''Predictions and labels have mismatched lengths {len(lowercase__ )} and {len(lowercase__ )}''' ) if task_name == "xnli": return {"acc": simple_accuracy(lowercase__ , lowercase__ )} else: raise KeyError(lowercase__ )
96
"""simple docstring""" import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler lowercase__ = 16 lowercase__ = 32 def _snake_case ( lowercase__ , lowercase__ = 16 , lowercase__ = "bert-base-cased" ): _lowerCamelCase : List[Any] = AutoTokenizer.from_pretrained(lowercase__ ) _lowerCamelCase : Tuple = load_dataset('glue' , 'mrpc' ) def tokenize_function(lowercase__ ): # max_length=None => use the model max length (it's actually the default) _lowerCamelCase : Union[str, Any] = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=lowercase__ , max_length=lowercase__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset _lowerCamelCase : int = datasets.map( lowercase__ , batched=lowercase__ , remove_columns=['idx', 'sentence1', 'sentence2'] , load_from_cache_file=lowercase__ ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _lowerCamelCase : Optional[int] = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(lowercase__ ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(lowercase__ , padding='max_length' , max_length=128 , return_tensors='pt' ) return tokenizer.pad(lowercase__ , padding='longest' , return_tensors='pt' ) # Instantiate dataloaders. _lowerCamelCase : List[str] = DataLoader( tokenized_datasets['train'] , shuffle=lowercase__ , collate_fn=lowercase__ , batch_size=lowercase__ ) _lowerCamelCase : int = DataLoader( tokenized_datasets['validation'] , shuffle=lowercase__ , collate_fn=lowercase__ , batch_size=lowercase__ ) return train_dataloader, eval_dataloader def _snake_case ( lowercase__ , lowercase__ ): # Initialize accelerator _lowerCamelCase : Optional[int] = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _lowerCamelCase : Optional[int] = config['lr'] _lowerCamelCase : Optional[int] = int(config['num_epochs'] ) _lowerCamelCase : Union[str, Any] = int(config['seed'] ) _lowerCamelCase : Optional[int] = int(config['batch_size'] ) _lowerCamelCase : Dict = args.model_name_or_path set_seed(lowercase__ ) _lowerCamelCase, _lowerCamelCase : Optional[int] = get_dataloaders(lowercase__ , lowercase__ , lowercase__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _lowerCamelCase : int = AutoModelForSequenceClassification.from_pretrained(lowercase__ , return_dict=lowercase__ ) # Instantiate optimizer _lowerCamelCase : Optional[int] = ( AdamW if accelerator.state.deepspeed_plugin is None or 'optimizer' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) _lowerCamelCase : Union[str, Any] = optimizer_cls(params=model.parameters() , lr=lowercase__ ) if accelerator.state.deepspeed_plugin is not None: _lowerCamelCase : str = accelerator.state.deepspeed_plugin.deepspeed_config[ 'gradient_accumulation_steps' ] else: _lowerCamelCase : Tuple = 1 _lowerCamelCase : List[Any] = (len(lowercase__ ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): _lowerCamelCase : Tuple = get_linear_schedule_with_warmup( optimizer=lowercase__ , num_warmup_steps=0 , num_training_steps=lowercase__ , ) else: _lowerCamelCase : Any = DummyScheduler(lowercase__ , total_num_steps=lowercase__ , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : Union[str, Any] = accelerator.prepare( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) # We need to keep track of how many total steps we have iterated over _lowerCamelCase : Union[str, Any] = 0 # We also need to keep track of the stating epoch so files are named properly _lowerCamelCase : Dict = 0 # Now we train the model _lowerCamelCase : Dict = evaluate.load('glue' , 'mrpc' ) _lowerCamelCase : Optional[int] = 0 _lowerCamelCase : str = {} for epoch in range(lowercase__ , lowercase__ ): model.train() for step, batch in enumerate(lowercase__ ): _lowerCamelCase : List[Any] = model(**lowercase__ ) _lowerCamelCase : int = outputs.loss _lowerCamelCase : Dict = loss / gradient_accumulation_steps accelerator.backward(lowercase__ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 model.eval() _lowerCamelCase : Union[str, Any] = 0 for step, batch in enumerate(lowercase__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _lowerCamelCase : Optional[int] = model(**lowercase__ ) _lowerCamelCase : Dict = outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times _lowerCamelCase, _lowerCamelCase : List[str] = accelerator.gather( (predictions, batch['labels']) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(lowercase__ ) - 1: _lowerCamelCase : Optional[Any] = predictions[: len(eval_dataloader.dataset ) - samples_seen] _lowerCamelCase : Dict = references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=lowercase__ , references=lowercase__ , ) _lowerCamelCase : List[Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f'''epoch {epoch}:''' , lowercase__ ) _lowerCamelCase : Tuple = eval_metric['accuracy'] if best_performance < eval_metric["accuracy"]: _lowerCamelCase : str = eval_metric['accuracy'] if args.performance_lower_bound is not None: assert ( args.performance_lower_bound <= best_performance ), f'''Best performance metric {best_performance} is lower than the lower bound {args.performance_lower_bound}''' accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , 'all_results.json' ) , 'w' ) as f: json.dump(lowercase__ , lowercase__ ) def _snake_case ( ): _lowerCamelCase : Any = argparse.ArgumentParser(description='Simple example of training script tracking peak GPU memory usage.' ) parser.add_argument( '--model_name_or_path' , type=lowercase__ , default='bert-base-cased' , help='Path to pretrained model or model identifier from huggingface.co/models.' , required=lowercase__ , ) parser.add_argument( '--output_dir' , type=lowercase__ , default='.' , help='Optional save directory where all checkpoint folders will be stored. Default is the current working directory.' , ) parser.add_argument( '--performance_lower_bound' , type=lowercase__ , default=lowercase__ , help='Optional lower bound for the performance metric. If set, the training will throw error when the performance metric drops below this value.' , ) parser.add_argument( '--num_epochs' , type=lowercase__ , default=3 , help='Number of train epochs.' , ) _lowerCamelCase : Optional[Any] = parser.parse_args() _lowerCamelCase : str = {'lr': 2E-5, 'num_epochs': args.num_epochs, 'seed': 42, 'batch_size': 16} training_function(lowercase__ , lowercase__ ) if __name__ == "__main__": main()
96
1
"""simple docstring""" # Lint as: python3 import os import re import urllib.parse from pathlib import Path from typing import Callable, List, Optional, Union from zipfile import ZipFile from ..utils.file_utils import cached_path, hf_github_url from ..utils.logging import get_logger from ..utils.version import Version lowercase__ = get_logger(__name__) class lowerCAmelCase__ : '''simple docstring''' lowerCamelCase__ = """dummy_data""" lowerCamelCase__ = """datasets""" lowerCamelCase__ = False def __init__( self , lowercase , lowercase , lowercase , lowercase = None , lowercase = False , lowercase = True , lowercase = None , ): _lowerCamelCase : Optional[Any] = 0 _lowerCamelCase : Dict = dataset_name _lowerCamelCase : Union[str, Any] = cache_dir _lowerCamelCase : Dict = use_local_dummy_data _lowerCamelCase : Tuple = config # download_callbacks take a single url as input _lowerCamelCase : List[Callable] = download_callbacks or [] # if False, it doesn't load existing files and it returns the paths of the dummy files relative # to the dummy_data zip file root _lowerCamelCase : Any = load_existing_dummy_data # TODO(PVP, QL) might need to make this more general _lowerCamelCase : str = str(lowercase ) # to be downloaded _lowerCamelCase : Union[str, Any] = None _lowerCamelCase : int = None @property def A_ ( self ): if self._dummy_file is None: _lowerCamelCase : Tuple = self.download_dummy_data() return self._dummy_file @property def A_ ( self ): if self.config is not None: # structure is dummy / config_name / version_name return os.path.join('dummy' , self.config.name , self.version_name ) # structure is dummy / version_name return os.path.join('dummy' , self.version_name ) @property def A_ ( self ): return os.path.join(self.dummy_data_folder , 'dummy_data.zip' ) def A_ ( self ): _lowerCamelCase : List[str] = ( self.local_path_to_dummy_data if self.use_local_dummy_data is True else self.github_path_to_dummy_data ) _lowerCamelCase : int = cached_path( lowercase , cache_dir=self.cache_dir , extract_compressed_file=lowercase , force_extract=lowercase ) return os.path.join(lowercase , self.dummy_file_name ) @property def A_ ( self ): return os.path.join(self.datasets_scripts_dir , self.dataset_name , self.dummy_zip_file ) @property def A_ ( self ): if self._bucket_url is None: _lowerCamelCase : List[Any] = hf_github_url(self.dataset_name , self.dummy_zip_file.replace(os.sep , '/' ) ) return self._bucket_url @property def A_ ( self ): # return full path if its a dir if os.path.isdir(self.dummy_file ): return self.dummy_file # else cut off path to file -> example `xsum`. return "/".join(self.dummy_file.replace(os.sep , '/' ).split('/' )[:-1] ) def A_ ( self , lowercase , *lowercase ): if self.load_existing_dummy_data: # dummy data is downloaded and tested _lowerCamelCase : Union[str, Any] = self.dummy_file else: # dummy data cannot be downloaded and only the path to dummy file is returned _lowerCamelCase : Union[str, Any] = self.dummy_file_name # special case when data_url is a dict if isinstance(lowercase , lowercase ): return self.create_dummy_data_dict(lowercase , lowercase ) elif isinstance(lowercase , (list, tuple) ): return self.create_dummy_data_list(lowercase , lowercase ) else: return self.create_dummy_data_single(lowercase , lowercase ) def A_ ( self , lowercase , *lowercase ): return self.download_and_extract(lowercase ) def A_ ( self , lowercase , lowercase ): return self.download_and_extract(lowercase ) def A_ ( self , lowercase , *lowercase , **lowercase ): return path def A_ ( self ): return {} def A_ ( self , lowercase , lowercase ): _lowerCamelCase : Optional[int] = {} for key, single_urls in data_url.items(): for download_callback in self.download_callbacks: if isinstance(lowercase , lowercase ): for single_url in single_urls: download_callback(lowercase ) else: _lowerCamelCase : List[Any] = single_urls download_callback(lowercase ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus if isinstance(lowercase , lowercase ): _lowerCamelCase : List[Any] = [os.path.join(lowercase , urllib.parse.quote_plus(Path(lowercase ).name ) ) for x in single_urls] else: _lowerCamelCase : Optional[int] = single_urls _lowerCamelCase : List[Any] = os.path.join(lowercase , urllib.parse.quote_plus(Path(lowercase ).name ) ) _lowerCamelCase : int = value # make sure that values are unique if all(isinstance(lowercase , lowercase ) for i in dummy_data_dict.values() ) and len(set(dummy_data_dict.values() ) ) < len( dummy_data_dict.values() ): # append key to value to make its name unique _lowerCamelCase : List[Any] = {key: value + key for key, value in dummy_data_dict.items()} return dummy_data_dict def A_ ( self , lowercase , lowercase ): _lowerCamelCase : Optional[Any] = [] # trick: if there are many shards named like `data.txt-000001-of-00300`, only use the first one _lowerCamelCase : List[str] = all(bool(re.findall('[0-9]{3,}-of-[0-9]{3,}' , lowercase ) ) for url in data_url ) _lowerCamelCase : int = all( url.startswith('https://ftp.ncbi.nlm.nih.gov/pubmed/baseline/pubmed' ) for url in data_url ) if data_url and (is_tf_records or is_pubmed_records): _lowerCamelCase : List[str] = [data_url[0]] * len(lowercase ) for single_url in data_url: for download_callback in self.download_callbacks: download_callback(lowercase ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus _lowerCamelCase : str = os.path.join(lowercase , urllib.parse.quote_plus(single_url.split('/' )[-1] ) ) dummy_data_list.append(lowercase ) return dummy_data_list def A_ ( self , lowercase , lowercase ): for download_callback in self.download_callbacks: download_callback(lowercase ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus _lowerCamelCase : Tuple = os.path.join(lowercase , urllib.parse.quote_plus(data_url.split('/' )[-1] ) ) if os.path.exists(lowercase ) or not self.load_existing_dummy_data: return value else: # Backward compatibility, maybe deprecate at one point. # For many datasets with single url calls to dl_manager.download_and_extract, # the dummy_data.zip file is actually the zipped downloaded file # while now we expected the dummy_data.zip file to be a directory containing # the downloaded file. return path_to_dummy_data def A_ ( self ): pass def A_ ( self ): pass def A_ ( self , lowercase ): def _iter_archive_members(lowercase ): # this preserves the order of the members inside the ZIP archive _lowerCamelCase : str = Path(self.dummy_file ).parent _lowerCamelCase : Union[str, Any] = path.relative_to(lowercase ) with ZipFile(self.local_path_to_dummy_data ) as zip_file: _lowerCamelCase : List[str] = zip_file.namelist() for member in members: if member.startswith(relative_path.as_posix() ): yield dummy_parent_path.joinpath(lowercase ) _lowerCamelCase : Optional[int] = Path(lowercase ) _lowerCamelCase : Dict = _iter_archive_members(lowercase ) if self.use_local_dummy_data else path.rglob('*' ) for file_path in file_paths: if file_path.is_file() and not file_path.name.startswith(('.', '__') ): yield file_path.relative_to(lowercase ).as_posix(), file_path.open('rb' ) def A_ ( self , lowercase ): if not isinstance(lowercase , lowercase ): _lowerCamelCase : List[str] = [paths] for path in paths: if os.path.isfile(lowercase ): if os.path.basename(lowercase ).startswith(('.', '__') ): return yield path else: for dirpath, dirnames, filenames in os.walk(lowercase ): if os.path.basename(lowercase ).startswith(('.', '__') ): continue dirnames.sort() for filename in sorted(lowercase ): if filename.startswith(('.', '__') ): continue yield os.path.join(lowercase , lowercase )
96
"""simple docstring""" from __future__ import annotations import copy import tempfile import unittest from transformers import CONFIG_MAPPING, AutoConfig, BertConfig, GPTaConfig, TaConfig, TapasConfig, is_tf_available from transformers.testing_utils import ( DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tensorflow_probability, require_tf, slow, ) from ..bert.test_modeling_bert import BertModelTester if is_tf_available(): from transformers import ( TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelForTableQuestionAnswering, TFAutoModelForTokenClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFFunnelBaseModel, TFFunnelModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, TFTapasForQuestionAnswering, ) from transformers.models.auto.modeling_tf_auto import ( TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_MAPPING, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.tapas.modeling_tf_tapas import TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """new-model""" if is_tf_available(): class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = NewModelConfig @require_tf class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' @slow def A_ ( self ): _lowerCamelCase : List[str] = 'bert-base-cased' _lowerCamelCase : Union[str, Any] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Union[str, Any] = TFAutoModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): _lowerCamelCase : List[str] = 'bert-base-cased' _lowerCamelCase : int = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : int = TFAutoModelForPreTraining.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Dict = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : int = TFAutoModelForCausalLM.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : str = TFAutoModelForCausalLM.from_pretrained(lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : List[Any] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : str = TFAutoModelWithLMHead.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Tuple = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Optional[int] = TFAutoModelForMaskedLM.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : Tuple = TFAutoModelForMaskedLM.from_pretrained(lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Optional[int] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Tuple = TFAutoModelForSeqaSeqLM.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : Tuple = TFAutoModelForSeqaSeqLM.from_pretrained(lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: _lowerCamelCase : str = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Union[str, Any] = TFAutoModelForSequenceClassification.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: _lowerCamelCase : Optional[Any] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : List[str] = TFAutoModelForQuestionAnswering.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow @require_tensorflow_probability def A_ ( self ): for model_name in TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST[5:6]: _lowerCamelCase : Dict = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Dict = TFAutoModelForTableQuestionAnswering.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : List[Any] = TFAutoModelForTableQuestionAnswering.from_pretrained( lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) def A_ ( self ): _lowerCamelCase : int = TFAutoModelWithLMHead.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowercase ) , 14410 ) def A_ ( self ): _lowerCamelCase : Any = TFAutoModelWithLMHead.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowercase ) , 14410 ) def A_ ( self ): # For the auto model mapping, FunnelConfig has two models: FunnelModel and FunnelBaseModel _lowerCamelCase : List[str] = TFAutoModel.from_pretrained('sgugger/funnel-random-tiny' ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Optional[int] = copy.deepcopy(model.config ) _lowerCamelCase : Dict = ['FunnelBaseModel'] _lowerCamelCase : List[Any] = TFAutoModel.from_config(lowercase ) self.assertIsInstance(lowercase , lowercase ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowercase ) _lowerCamelCase : Tuple = TFAutoModel.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) def A_ ( self ): try: AutoConfig.register('new-model' , lowercase ) _lowerCamelCase : Tuple = [ TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSequenceClassification, TFAutoModelForTokenClassification, ] for auto_class in auto_classes: with self.subTest(auto_class.__name__ ): # Wrong config class will raise an error with self.assertRaises(lowercase ): auto_class.register(lowercase , lowercase ) auto_class.register(lowercase , lowercase ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(lowercase ): auto_class.register(lowercase , lowercase ) # Now that the config is registered, it can be used as any other config with the auto-API _lowerCamelCase : Optional[Any] = BertModelTester(self ).get_config() _lowerCamelCase : Dict = NewModelConfig(**tiny_config.to_dict() ) _lowerCamelCase : int = auto_class.from_config(lowercase ) self.assertIsInstance(lowercase , lowercase ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowercase ) _lowerCamelCase : List[Any] = auto_class.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"] for mapping in ( TF_MODEL_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, ): if NewModelConfig in mapping._extra_content: del mapping._extra_content[NewModelConfig] def A_ ( self ): with self.assertRaisesRegex( lowercase , 'bert-base is not a local folder and is not a valid model identifier' ): _lowerCamelCase : Union[str, Any] = TFAutoModel.from_pretrained('bert-base' ) def A_ ( self ): with self.assertRaisesRegex( lowercase , r'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): _lowerCamelCase : str = TFAutoModel.from_pretrained(lowercase , revision='aaaaaa' ) def A_ ( self ): with self.assertRaisesRegex( lowercase , 'hf-internal-testing/config-no-model does not appear to have a file named pytorch_model.bin' , ): _lowerCamelCase : Tuple = TFAutoModel.from_pretrained('hf-internal-testing/config-no-model' ) def A_ ( self ): with self.assertRaisesRegex(lowercase , 'Use `from_pt=True` to load this model' ): _lowerCamelCase : Tuple = TFAutoModel.from_pretrained('hf-internal-testing/tiny-bert-pt-only' ) def A_ ( self ): # Make sure we have cached the model. _lowerCamelCase : Optional[int] = TFAutoModel.from_pretrained('hf-internal-testing/tiny-random-bert' ) with RequestCounter() as counter: _lowerCamelCase : Optional[int] = TFAutoModel.from_pretrained('hf-internal-testing/tiny-random-bert' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 ) # With a sharded checkpoint _lowerCamelCase : int = TFAutoModel.from_pretrained('ArthurZ/tiny-random-bert-sharded' ) with RequestCounter() as counter: _lowerCamelCase : List[Any] = TFAutoModel.from_pretrained('ArthurZ/tiny-random-bert-sharded' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 )
96
1
"""simple docstring""" from __future__ import annotations import math import numpy as np from numpy.linalg import norm def _snake_case ( lowercase__ , lowercase__ ): return math.sqrt(sum(pow(a - b , 2 ) for a, b in zip(lowercase__ , lowercase__ ) ) ) def _snake_case ( lowercase__ , lowercase__ ): if dataset.ndim != value_array.ndim: _lowerCamelCase : Tuple = ( 'Wrong input data\'s dimensions... ' f'''dataset : {dataset.ndim}, value_array : {value_array.ndim}''' ) raise ValueError(lowercase__ ) try: if dataset.shape[1] != value_array.shape[1]: _lowerCamelCase : Optional[int] = ( 'Wrong input data\'s shape... ' f'''dataset : {dataset.shape[1]}, value_array : {value_array.shape[1]}''' ) raise ValueError(lowercase__ ) except IndexError: if dataset.ndim != value_array.ndim: raise TypeError('Wrong shape' ) if dataset.dtype != value_array.dtype: _lowerCamelCase : int = ( 'Input data have different datatype... ' f'''dataset : {dataset.dtype}, value_array : {value_array.dtype}''' ) raise TypeError(lowercase__ ) _lowerCamelCase : Optional[int] = [] for value in value_array: _lowerCamelCase : Tuple = euclidean(lowercase__ , dataset[0] ) _lowerCamelCase : Union[str, Any] = dataset[0].tolist() for dataset_value in dataset[1:]: _lowerCamelCase : Optional[Any] = euclidean(lowercase__ , lowercase__ ) if dist > temp_dist: _lowerCamelCase : List[Any] = temp_dist _lowerCamelCase : List[str] = dataset_value.tolist() answer.append([vector, dist] ) return answer def _snake_case ( lowercase__ , lowercase__ ): return np.dot(lowercase__ , lowercase__ ) / (norm(lowercase__ ) * norm(lowercase__ )) if __name__ == "__main__": import doctest doctest.testmod()
96
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ = {"""configuration_ibert""": ["""IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """IBertConfig""", """IBertOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ """IBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """IBertForMaskedLM""", """IBertForMultipleChoice""", """IBertForQuestionAnswering""", """IBertForSequenceClassification""", """IBertForTokenClassification""", """IBertModel""", """IBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
96
1
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import warnings from typing import List from unittest.mock import Mock import torch from torch.utils.data import DataLoader, IterableDataset, TensorDataset from accelerate.accelerator import Accelerator from accelerate.utils.dataclasses import DistributedType class lowerCAmelCase__ ( lowercase ): '''simple docstring''' def __init__( self , lowercase ): _lowerCamelCase : Dict = data def __iter__( self ): for element in self.data: yield element def _snake_case ( lowercase__=True ): _lowerCamelCase : Any = Accelerator(even_batches=lowercase__ ) assert accelerator.num_processes == 2, "this script expects that two GPUs are available" return accelerator def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ = False ): if iterable: _lowerCamelCase : List[Any] = DummyIterableDataset(torch.as_tensor(range(lowercase__ ) ) ) else: _lowerCamelCase : List[str] = TensorDataset(torch.as_tensor(range(lowercase__ ) ) ) _lowerCamelCase : Any = DataLoader(lowercase__ , batch_size=lowercase__ ) _lowerCamelCase : str = accelerator.prepare(lowercase__ ) return dl def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , ): _lowerCamelCase : int = create_dataloader(accelerator=lowercase__ , dataset_size=lowercase__ , batch_size=lowercase__ ) _lowerCamelCase : Optional[Any] = [len(batch[0] ) for batch in dl] if accelerator.process_index == 0: assert batch_sizes == process_0_expected_batch_sizes elif accelerator.process_index == 1: assert batch_sizes == process_1_expected_batch_sizes def _snake_case ( ): _lowerCamelCase : int = create_accelerator() # without padding, we would expect a different number of batches verify_dataloader_batch_sizes( lowercase__ , dataset_size=3 , batch_size=1 , process_0_expected_batch_sizes=[1, 1] , process_1_expected_batch_sizes=[1, 1] , ) # without padding, we would expect the same number of batches, but different sizes verify_dataloader_batch_sizes( lowercase__ , dataset_size=7 , batch_size=2 , process_0_expected_batch_sizes=[2, 2] , process_1_expected_batch_sizes=[2, 2] , ) def _snake_case ( ): _lowerCamelCase : Optional[int] = create_accelerator(even_batches=lowercase__ ) verify_dataloader_batch_sizes( lowercase__ , dataset_size=3 , batch_size=1 , process_0_expected_batch_sizes=[1, 1] , process_1_expected_batch_sizes=[1] , ) verify_dataloader_batch_sizes( lowercase__ , dataset_size=7 , batch_size=2 , process_0_expected_batch_sizes=[2, 2] , process_1_expected_batch_sizes=[2, 1] , ) def _snake_case ( ): _lowerCamelCase : int = create_accelerator(even_batches=lowercase__ ) _lowerCamelCase : str = torch.nn.Linear(1 , 1 ) _lowerCamelCase : Dict = accelerator.prepare(lowercase__ ) _lowerCamelCase : Optional[Any] = create_dataloader(lowercase__ , dataset_size=3 , batch_size=1 ) _lowerCamelCase : Optional[Any] = [] with accelerator.join_uneven_inputs([ddp_model] ): for batch_idx, batch in enumerate(lowercase__ ): _lowerCamelCase : str = ddp_model(batch[0].float() ) _lowerCamelCase : str = output.sum() loss.backward() batch_idxs.append(lowercase__ ) accelerator.wait_for_everyone() if accelerator.process_index == 0: assert batch_idxs == [0, 1] elif accelerator.process_index == 1: assert batch_idxs == [0] def _snake_case ( lowercase__ ): with warnings.catch_warnings(record=lowercase__ ) as w: with accelerator.join_uneven_inputs([Mock()] ): pass assert issubclass(w[-1].category , lowercase__ ) assert "only supported for multi-GPU" in str(w[-1].message ) def _snake_case ( ): _lowerCamelCase : Any = True _lowerCamelCase : Any = False _lowerCamelCase : List[str] = create_accelerator(even_batches=lowercase__ ) _lowerCamelCase : Dict = torch.nn.Linear(1 , 1 ) _lowerCamelCase : Union[str, Any] = accelerator.prepare(lowercase__ ) _lowerCamelCase : Any = create_dataloader(lowercase__ , dataset_size=3 , batch_size=1 ) _lowerCamelCase : Optional[Any] = create_dataloader(lowercase__ , dataset_size=3 , batch_size=1 ) with accelerator.join_uneven_inputs([ddp_model] , even_batches=lowercase__ ): _lowerCamelCase : Dict = train_dl.batch_sampler.even_batches _lowerCamelCase : Union[str, Any] = valid_dl.batch_sampler.even_batches assert train_dl_overridden_value == overridden_even_batches assert valid_dl_overridden_value == overridden_even_batches assert train_dl.batch_sampler.even_batches == default_even_batches assert valid_dl.batch_sampler.even_batches == default_even_batches def _snake_case ( ): _lowerCamelCase : Dict = True _lowerCamelCase : Tuple = False _lowerCamelCase : Dict = create_accelerator(even_batches=lowercase__ ) _lowerCamelCase : Dict = torch.nn.Linear(1 , 1 ) _lowerCamelCase : Union[str, Any] = accelerator.prepare(lowercase__ ) create_dataloader(lowercase__ , dataset_size=3 , batch_size=1 , iterable=lowercase__ ) _lowerCamelCase : str = create_dataloader(lowercase__ , dataset_size=3 , batch_size=1 ) with warnings.catch_warnings(): warnings.filterwarnings('ignore' ) try: with accelerator.join_uneven_inputs([ddp_model] , even_batches=lowercase__ ): _lowerCamelCase : Any = batch_dl.batch_sampler.even_batches except AttributeError: # ensure attribute error is not raised when processing iterable dl raise AssertionError assert batch_dl_overridden_value == overridden_even_batches assert batch_dl.batch_sampler.even_batches == default_even_batches def _snake_case ( ): _lowerCamelCase : List[Any] = create_accelerator() _lowerCamelCase : Any = torch.nn.Linear(1 , 1 ) _lowerCamelCase : Any = accelerator.prepare(lowercase__ ) create_dataloader(lowercase__ , dataset_size=3 , batch_size=1 , iterable=lowercase__ ) with warnings.catch_warnings(record=lowercase__ ) as w: with accelerator.join_uneven_inputs([ddp_model] , even_batches=lowercase__ ): pass assert issubclass(w[-1].category , lowercase__ ) assert "only supported for map-style datasets" in str(w[-1].message ) def _snake_case ( ): _lowerCamelCase : Optional[int] = create_accelerator() accelerator.print('Test that even_batches variable ensures uniform batches across processes' ) test_default_ensures_even_batch_sizes() accelerator.print('Run tests with even_batches disabled' ) test_can_disable_even_batches() accelerator.print('Test joining uneven inputs' ) test_can_join_uneven_inputs() accelerator.print('Test overriding even_batches when joining uneven inputs' ) test_join_can_override_even_batches() accelerator.print('Test overriding even_batches for mixed dataloader types' ) test_join_can_override_for_mixed_type_dataloaders() accelerator.print('Test overriding even_batches raises a warning for iterable dataloaders' ) test_join_raises_warning_for_iterable_when_overriding_even_batches() accelerator.print('Test join with non DDP distributed raises warning' ) _lowerCamelCase : Dict = accelerator.state.distributed_type _lowerCamelCase : List[Any] = DistributedType.FSDP test_join_raises_warning_for_non_ddp_distributed(lowercase__ ) _lowerCamelCase : Union[str, Any] = original_state if __name__ == "__main__": main()
96
"""simple docstring""" import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : Tuple = f'''{sampling_rate}''' _lowerCamelCase : str = '1' _lowerCamelCase : str = 'f32le' _lowerCamelCase : Union[str, Any] = [ 'ffmpeg', '-i', 'pipe:0', '-ac', ac, '-ar', ar, '-f', format_for_conversion, '-hide_banner', '-loglevel', 'quiet', 'pipe:1', ] try: with subprocess.Popen(lowercase__ , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: _lowerCamelCase : str = ffmpeg_process.communicate(lowercase__ ) except FileNotFoundError as error: raise ValueError('ffmpeg was not found but is required to load audio files from filename' ) from error _lowerCamelCase : List[Any] = output_stream[0] _lowerCamelCase : Tuple = np.frombuffer(lowercase__ , np.floataa ) if audio.shape[0] == 0: raise ValueError('Malformed soundfile' ) return audio def _snake_case ( lowercase__ , lowercase__ , lowercase__ = "f32le" , ): _lowerCamelCase : Optional[Any] = f'''{sampling_rate}''' _lowerCamelCase : List[str] = '1' if format_for_conversion == "s16le": _lowerCamelCase : List[str] = 2 elif format_for_conversion == "f32le": _lowerCamelCase : List[Any] = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) _lowerCamelCase : Dict = platform.system() if system == "Linux": _lowerCamelCase : Optional[int] = 'alsa' _lowerCamelCase : Optional[Any] = 'default' elif system == "Darwin": _lowerCamelCase : Optional[int] = 'avfoundation' _lowerCamelCase : Any = ':0' elif system == "Windows": _lowerCamelCase : Tuple = 'dshow' _lowerCamelCase : Tuple = 'default' _lowerCamelCase : Optional[int] = [ 'ffmpeg', '-f', format_, '-i', input_, '-ac', ac, '-ar', ar, '-f', format_for_conversion, '-fflags', 'nobuffer', '-hide_banner', '-loglevel', 'quiet', 'pipe:1', ] _lowerCamelCase : Tuple = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample _lowerCamelCase : List[Any] = _ffmpeg_stream(lowercase__ , lowercase__ ) for item in iterator: yield item def _snake_case ( lowercase__ , lowercase__ , lowercase__ = None , lowercase__ = None , lowercase__ = "f32le" , ): if stream_chunk_s is not None: _lowerCamelCase : int = stream_chunk_s else: _lowerCamelCase : Optional[Any] = chunk_length_s _lowerCamelCase : Optional[Any] = ffmpeg_microphone(lowercase__ , lowercase__ , format_for_conversion=lowercase__ ) if format_for_conversion == "s16le": _lowerCamelCase : List[str] = np.intaa _lowerCamelCase : str = 2 elif format_for_conversion == "f32le": _lowerCamelCase : Any = np.floataa _lowerCamelCase : List[Any] = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) if stride_length_s is None: _lowerCamelCase : Union[str, Any] = chunk_length_s / 6 _lowerCamelCase : Optional[int] = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(lowercase__ , (int, float) ): _lowerCamelCase : Any = [stride_length_s, stride_length_s] _lowerCamelCase : Tuple = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample _lowerCamelCase : Optional[Any] = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample _lowerCamelCase : List[Any] = datetime.datetime.now() _lowerCamelCase : Optional[int] = datetime.timedelta(seconds=lowercase__ ) for item in chunk_bytes_iter(lowercase__ , lowercase__ , stride=(stride_left, stride_right) , stream=lowercase__ ): # Put everything back in numpy scale _lowerCamelCase : List[Any] = np.frombuffer(item['raw'] , dtype=lowercase__ ) _lowerCamelCase : int = ( item['stride'][0] // size_of_sample, item['stride'][1] // size_of_sample, ) _lowerCamelCase : Optional[int] = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ = False ): _lowerCamelCase : int = B'' _lowerCamelCase, _lowerCamelCase : Dict = stride if stride_left + stride_right >= chunk_len: raise ValueError( f'''Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}''' ) _lowerCamelCase : str = 0 for raw in iterator: acc += raw if stream and len(lowercase__ ) < chunk_len: _lowerCamelCase : Optional[int] = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(lowercase__ ) >= chunk_len: # We are flushing the accumulator _lowerCamelCase : str = (_stride_left, stride_right) _lowerCamelCase : str = {'raw': acc[:chunk_len], 'stride': stride} if stream: _lowerCamelCase : List[Any] = False yield item _lowerCamelCase : Optional[Any] = stride_left _lowerCamelCase : str = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(lowercase__ ) > stride_left: _lowerCamelCase : Optional[Any] = {'raw': acc, 'stride': (_stride_left, 0)} if stream: _lowerCamelCase : Tuple = False yield item def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : int = 2**24 # 16Mo try: with subprocess.Popen(lowercase__ , stdout=subprocess.PIPE , bufsize=lowercase__ ) as ffmpeg_process: while True: _lowerCamelCase : Optional[Any] = ffmpeg_process.stdout.read(lowercase__ ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError('ffmpeg was not found but is required to stream audio files from filename' ) from error
96
1
"""simple docstring""" import math class lowerCAmelCase__ : '''simple docstring''' def __init__( self , lowercase=0 ): # a graph with Node 0,1,...,N-1 _lowerCamelCase : List[str] = n _lowerCamelCase : str = [ [math.inf for j in range(0 , lowercase )] for i in range(0 , lowercase ) ] # adjacency matrix for weight _lowerCamelCase : List[str] = [ [math.inf for j in range(0 , lowercase )] for i in range(0 , lowercase ) ] # dp[i][j] stores minimum distance from i to j def A_ ( self , lowercase , lowercase , lowercase ): _lowerCamelCase : Union[str, Any] = w def A_ ( self ): for k in range(0 , self.n ): for i in range(0 , self.n ): for j in range(0 , self.n ): _lowerCamelCase : Dict = min(self.dp[i][j] , self.dp[i][k] + self.dp[k][j] ) def A_ ( self , lowercase , lowercase ): return self.dp[u][v] if __name__ == "__main__": lowercase__ = Graph(5) graph.add_edge(0, 2, 9) graph.add_edge(0, 4, 10) graph.add_edge(1, 3, 5) graph.add_edge(2, 3, 7) graph.add_edge(3, 0, 10) graph.add_edge(3, 1, 2) graph.add_edge(3, 2, 1) graph.add_edge(3, 4, 6) graph.add_edge(4, 1, 3) graph.add_edge(4, 2, 4) graph.add_edge(4, 3, 9) graph.floyd_warshall() graph.show_min(1, 4) graph.show_min(0, 3)
96
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = {"""ctrl""": """https://huggingface.co/ctrl/resolve/main/config.json"""} class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """ctrl""" lowerCamelCase__ = ["""past_key_values"""] lowerCamelCase__ = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowercase=246534 , lowercase=256 , lowercase=1280 , lowercase=8192 , lowercase=48 , lowercase=16 , lowercase=0.1 , lowercase=0.1 , lowercase=1E-6 , lowercase=0.02 , lowercase=True , **lowercase , ): _lowerCamelCase : Any = vocab_size _lowerCamelCase : Dict = n_positions _lowerCamelCase : Optional[int] = n_embd _lowerCamelCase : str = n_layer _lowerCamelCase : Union[str, Any] = n_head _lowerCamelCase : Any = dff _lowerCamelCase : int = resid_pdrop _lowerCamelCase : Dict = embd_pdrop _lowerCamelCase : Union[str, Any] = layer_norm_epsilon _lowerCamelCase : Tuple = initializer_range _lowerCamelCase : str = use_cache super().__init__(**lowercase )
96
1
"""simple docstring""" import re import string from collections import Counter import sacrebleu import sacremoses from packaging import version import datasets lowercase__ = """ @inproceedings{xu-etal-2016-optimizing, title = {Optimizing Statistical Machine Translation for Text Simplification}, authors={Xu, Wei and Napoles, Courtney and Pavlick, Ellie and Chen, Quanze and Callison-Burch, Chris}, journal = {Transactions of the Association for Computational Linguistics}, volume = {4}, year={2016}, url = {https://www.aclweb.org/anthology/Q16-1029}, pages = {401--415 }, @inproceedings{post-2018-call, title = \"A Call for Clarity in Reporting {BLEU} Scores\", author = \"Post, Matt\", booktitle = \"Proceedings of the Third Conference on Machine Translation: Research Papers\", month = oct, year = \"2018\", address = \"Belgium, Brussels\", publisher = \"Association for Computational Linguistics\", url = \"https://www.aclweb.org/anthology/W18-6319\", pages = \"186--191\", } """ lowercase__ = """\ WIKI_SPLIT is the combination of three metrics SARI, EXACT and SACREBLEU It can be used to evaluate the quality of machine-generated texts. """ lowercase__ = """ Calculates sari score (between 0 and 100) given a list of source and predicted sentences, and a list of lists of reference sentences. It also computes the BLEU score as well as the exact match score. Args: sources: list of source sentences where each sentence should be a string. predictions: list of predicted sentences where each sentence should be a string. references: list of lists of reference sentences where each sentence should be a string. Returns: sari: sari score sacrebleu: sacrebleu score exact: exact score Examples: >>> sources=[\"About 95 species are currently accepted .\"] >>> predictions=[\"About 95 you now get in .\"] >>> references=[[\"About 95 species are currently known .\"]] >>> wiki_split = datasets.load_metric(\"wiki_split\") >>> results = wiki_split.compute(sources=sources, predictions=predictions, references=references) >>> print(results) {'sari': 21.805555555555557, 'sacrebleu': 14.535768424205482, 'exact': 0.0} """ def _snake_case ( lowercase__ ): def remove_articles(lowercase__ ): _lowerCamelCase : str = re.compile(r'\b(a|an|the)\b' , re.UNICODE ) return re.sub(lowercase__ , ' ' , lowercase__ ) def white_space_fix(lowercase__ ): return " ".join(text.split() ) def remove_punc(lowercase__ ): _lowerCamelCase : Tuple = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(lowercase__ ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(lowercase__ ) ) ) ) def _snake_case ( lowercase__ , lowercase__ ): return int(normalize_answer(lowercase__ ) == normalize_answer(lowercase__ ) ) def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : Optional[int] = [any(compute_exact(lowercase__ , lowercase__ ) for ref in refs ) for pred, refs in zip(lowercase__ , lowercase__ )] return (sum(lowercase__ ) / len(lowercase__ )) * 100 def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : Optional[int] = [rgram for rgrams in rgramslist for rgram in rgrams] _lowerCamelCase : str = Counter(lowercase__ ) _lowerCamelCase : Tuple = Counter(lowercase__ ) _lowerCamelCase : Any = Counter() for sgram, scount in sgramcounter.items(): _lowerCamelCase : Optional[Any] = scount * numref _lowerCamelCase : Tuple = Counter(lowercase__ ) _lowerCamelCase : List[Any] = Counter() for cgram, ccount in cgramcounter.items(): _lowerCamelCase : Tuple = ccount * numref # KEEP _lowerCamelCase : List[Any] = sgramcounter_rep & cgramcounter_rep _lowerCamelCase : int = keepgramcounter_rep & rgramcounter _lowerCamelCase : str = sgramcounter_rep & rgramcounter _lowerCamelCase : List[Any] = 0 _lowerCamelCase : Tuple = 0 for keepgram in keepgramcountergood_rep: keeptmpscorea += keepgramcountergood_rep[keepgram] / keepgramcounter_rep[keepgram] # Fix an alleged bug [2] in the keep score computation. # keeptmpscore2 += keepgramcountergood_rep[keepgram] / keepgramcounterall_rep[keepgram] keeptmpscorea += keepgramcountergood_rep[keepgram] # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. _lowerCamelCase : int = 1 _lowerCamelCase : int = 1 if len(lowercase__ ) > 0: _lowerCamelCase : Dict = keeptmpscorea / len(lowercase__ ) if len(lowercase__ ) > 0: # Fix an alleged bug [2] in the keep score computation. # keepscore_recall = keeptmpscore2 / len(keepgramcounterall_rep) _lowerCamelCase : List[Any] = keeptmpscorea / sum(keepgramcounterall_rep.values() ) _lowerCamelCase : str = 0 if keepscore_precision > 0 or keepscore_recall > 0: _lowerCamelCase : Dict = 2 * keepscore_precision * keepscore_recall / (keepscore_precision + keepscore_recall) # DELETION _lowerCamelCase : Optional[Any] = sgramcounter_rep - cgramcounter_rep _lowerCamelCase : List[Any] = delgramcounter_rep - rgramcounter _lowerCamelCase : List[str] = sgramcounter_rep - rgramcounter _lowerCamelCase : str = 0 _lowerCamelCase : Dict = 0 for delgram in delgramcountergood_rep: deltmpscorea += delgramcountergood_rep[delgram] / delgramcounter_rep[delgram] deltmpscorea += delgramcountergood_rep[delgram] / delgramcounterall_rep[delgram] # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. _lowerCamelCase : str = 1 if len(lowercase__ ) > 0: _lowerCamelCase : Dict = deltmpscorea / len(lowercase__ ) # ADDITION _lowerCamelCase : Union[str, Any] = set(lowercase__ ) - set(lowercase__ ) _lowerCamelCase : Tuple = set(lowercase__ ) & set(lowercase__ ) _lowerCamelCase : str = set(lowercase__ ) - set(lowercase__ ) _lowerCamelCase : Union[str, Any] = 0 for addgram in addgramcountergood: addtmpscore += 1 # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. _lowerCamelCase : int = 1 _lowerCamelCase : Union[str, Any] = 1 if len(lowercase__ ) > 0: _lowerCamelCase : Any = addtmpscore / len(lowercase__ ) if len(lowercase__ ) > 0: _lowerCamelCase : Any = addtmpscore / len(lowercase__ ) _lowerCamelCase : Union[str, Any] = 0 if addscore_precision > 0 or addscore_recall > 0: _lowerCamelCase : Union[str, Any] = 2 * addscore_precision * addscore_recall / (addscore_precision + addscore_recall) return (keepscore, delscore_precision, addscore) def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : Tuple = len(lowercase__ ) _lowerCamelCase : str = ssent.split(' ' ) _lowerCamelCase : List[Any] = csent.split(' ' ) _lowerCamelCase : Optional[Any] = [] _lowerCamelCase : Optional[int] = [] _lowerCamelCase : Dict = [] _lowerCamelCase : Optional[int] = [] _lowerCamelCase : List[Any] = [] _lowerCamelCase : Dict = [] _lowerCamelCase : Union[str, Any] = [] _lowerCamelCase : Dict = [] _lowerCamelCase : int = [] _lowerCamelCase : List[Any] = [] for rsent in rsents: _lowerCamelCase : Tuple = rsent.split(' ' ) _lowerCamelCase : Any = [] _lowerCamelCase : int = [] _lowerCamelCase : Tuple = [] ragramslist.append(lowercase__ ) for i in range(0 , len(lowercase__ ) - 1 ): if i < len(lowercase__ ) - 1: _lowerCamelCase : str = ragrams[i] + ' ' + ragrams[i + 1] ragrams.append(lowercase__ ) if i < len(lowercase__ ) - 2: _lowerCamelCase : List[str] = ragrams[i] + ' ' + ragrams[i + 1] + ' ' + ragrams[i + 2] ragrams.append(lowercase__ ) if i < len(lowercase__ ) - 3: _lowerCamelCase : List[str] = ragrams[i] + ' ' + ragrams[i + 1] + ' ' + ragrams[i + 2] + ' ' + ragrams[i + 3] ragrams.append(lowercase__ ) ragramslist.append(lowercase__ ) ragramslist.append(lowercase__ ) ragramslist.append(lowercase__ ) for i in range(0 , len(lowercase__ ) - 1 ): if i < len(lowercase__ ) - 1: _lowerCamelCase : List[str] = sagrams[i] + ' ' + sagrams[i + 1] sagrams.append(lowercase__ ) if i < len(lowercase__ ) - 2: _lowerCamelCase : List[Any] = sagrams[i] + ' ' + sagrams[i + 1] + ' ' + sagrams[i + 2] sagrams.append(lowercase__ ) if i < len(lowercase__ ) - 3: _lowerCamelCase : List[Any] = sagrams[i] + ' ' + sagrams[i + 1] + ' ' + sagrams[i + 2] + ' ' + sagrams[i + 3] sagrams.append(lowercase__ ) for i in range(0 , len(lowercase__ ) - 1 ): if i < len(lowercase__ ) - 1: _lowerCamelCase : Optional[int] = cagrams[i] + ' ' + cagrams[i + 1] cagrams.append(lowercase__ ) if i < len(lowercase__ ) - 2: _lowerCamelCase : Optional[Any] = cagrams[i] + ' ' + cagrams[i + 1] + ' ' + cagrams[i + 2] cagrams.append(lowercase__ ) if i < len(lowercase__ ) - 3: _lowerCamelCase : Optional[int] = cagrams[i] + ' ' + cagrams[i + 1] + ' ' + cagrams[i + 2] + ' ' + cagrams[i + 3] cagrams.append(lowercase__ ) ((_lowerCamelCase), (_lowerCamelCase), (_lowerCamelCase)) : Optional[Any] = SARIngram(lowercase__ , lowercase__ , lowercase__ , lowercase__ ) ((_lowerCamelCase), (_lowerCamelCase), (_lowerCamelCase)) : Tuple = SARIngram(lowercase__ , lowercase__ , lowercase__ , lowercase__ ) ((_lowerCamelCase), (_lowerCamelCase), (_lowerCamelCase)) : List[str] = SARIngram(lowercase__ , lowercase__ , lowercase__ , lowercase__ ) ((_lowerCamelCase), (_lowerCamelCase), (_lowerCamelCase)) : Optional[int] = SARIngram(lowercase__ , lowercase__ , lowercase__ , lowercase__ ) _lowerCamelCase : List[Any] = sum([keepascore, keepascore, keepascore, keepascore] ) / 4 _lowerCamelCase : int = sum([delascore, delascore, delascore, delascore] ) / 4 _lowerCamelCase : Union[str, Any] = sum([addascore, addascore, addascore, addascore] ) / 4 _lowerCamelCase : List[Any] = (avgkeepscore + avgdelscore + avgaddscore) / 3 return finalscore def _snake_case ( lowercase__ , lowercase__ = True , lowercase__ = "13a" , lowercase__ = True ): # Normalization is requried for the ASSET dataset (one of the primary # datasets in sentence simplification) to allow using space # to split the sentence. Even though Wiki-Auto and TURK datasets, # do not require normalization, we do it for consistency. # Code adapted from the EASSE library [1] written by the authors of the ASSET dataset. # [1] https://github.com/feralvam/easse/blob/580bba7e1378fc8289c663f864e0487188fe8067/easse/utils/preprocessing.py#L7 if lowercase: _lowerCamelCase : Union[str, Any] = sentence.lower() if tokenizer in ["13a", "intl"]: if version.parse(sacrebleu.__version__ ).major >= 2: _lowerCamelCase : Union[str, Any] = sacrebleu.metrics.bleu._get_tokenizer(lowercase__ )()(lowercase__ ) else: _lowerCamelCase : Optional[int] = sacrebleu.TOKENIZERS[tokenizer]()(lowercase__ ) elif tokenizer == "moses": _lowerCamelCase : str = sacremoses.MosesTokenizer().tokenize(lowercase__ , return_str=lowercase__ , escape=lowercase__ ) elif tokenizer == "penn": _lowerCamelCase : List[str] = sacremoses.MosesTokenizer().penn_tokenize(lowercase__ , return_str=lowercase__ ) else: _lowerCamelCase : str = sentence if not return_str: _lowerCamelCase : List[Any] = normalized_sent.split() return normalized_sent def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): if not (len(lowercase__ ) == len(lowercase__ ) == len(lowercase__ )): raise ValueError('Sources length must match predictions and references lengths.' ) _lowerCamelCase : List[str] = 0 for src, pred, refs in zip(lowercase__ , lowercase__ , lowercase__ ): sari_score += SARIsent(normalize(lowercase__ ) , normalize(lowercase__ ) , [normalize(lowercase__ ) for sent in refs] ) _lowerCamelCase : List[str] = sari_score / len(lowercase__ ) return 100 * sari_score def _snake_case ( lowercase__ , lowercase__ , lowercase__="exp" , lowercase__=None , lowercase__=False , lowercase__=False , lowercase__=False , ): _lowerCamelCase : Any = len(references[0] ) if any(len(lowercase__ ) != references_per_prediction for refs in references ): raise ValueError('Sacrebleu requires the same number of references for each prediction' ) _lowerCamelCase : Optional[Any] = [[refs[i] for refs in references] for i in range(lowercase__ )] _lowerCamelCase : Optional[Any] = sacrebleu.corpus_bleu( lowercase__ , lowercase__ , smooth_method=lowercase__ , smooth_value=lowercase__ , force=lowercase__ , lowercase=lowercase__ , use_effective_order=lowercase__ , ) return output.score @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCAmelCase__ ( datasets.Metric ): '''simple docstring''' def A_ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Sequence(datasets.Value('string' , id='sequence' ) , id='references' ), } ) , codebase_urls=[ 'https://github.com/huggingface/transformers/blob/master/src/transformers/data/metrics/squad_metrics.py', 'https://github.com/cocoxu/simplification/blob/master/SARI.py', 'https://github.com/tensorflow/tensor2tensor/blob/master/tensor2tensor/utils/sari_hook.py', 'https://github.com/mjpost/sacreBLEU', ] , reference_urls=[ 'https://www.aclweb.org/anthology/Q16-1029.pdf', 'https://github.com/mjpost/sacreBLEU', 'https://en.wikipedia.org/wiki/BLEU', 'https://towardsdatascience.com/evaluating-text-output-in-nlp-bleu-at-your-own-risk-e8609665a213', ] , ) def A_ ( self , lowercase , lowercase , lowercase ): _lowerCamelCase : Optional[int] = {} result.update({'sari': compute_sari(sources=lowercase , predictions=lowercase , references=lowercase )} ) result.update({'sacrebleu': compute_sacrebleu(predictions=lowercase , references=lowercase )} ) result.update({'exact': compute_em(predictions=lowercase , references=lowercase )} ) return result
96
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator from typing import Any class lowerCAmelCase__ : '''simple docstring''' def __init__( self , lowercase ): _lowerCamelCase : Any = data _lowerCamelCase : Node | None = None class lowerCAmelCase__ : '''simple docstring''' def __init__( self ): _lowerCamelCase : str = None _lowerCamelCase : str = None def __iter__( self ): _lowerCamelCase : List[str] = self.head while self.head: yield node.data _lowerCamelCase : Optional[int] = node.next if node == self.head: break def __len__( self ): return sum(1 for _ in self ) def __repr__( self ): return "->".join(str(lowercase ) for item in iter(self ) ) def A_ ( self , lowercase ): self.insert_nth(len(self ) , lowercase ) def A_ ( self , lowercase ): self.insert_nth(0 , lowercase ) def A_ ( self , lowercase , lowercase ): if index < 0 or index > len(self ): raise IndexError('list index out of range.' ) _lowerCamelCase : List[Any] = Node(lowercase ) if self.head is None: _lowerCamelCase : str = new_node # first node points itself _lowerCamelCase : Union[str, Any] = new_node elif index == 0: # insert at head _lowerCamelCase : List[str] = self.head _lowerCamelCase : str = new_node else: _lowerCamelCase : Union[str, Any] = self.head for _ in range(index - 1 ): _lowerCamelCase : List[Any] = temp.next _lowerCamelCase : Union[str, Any] = temp.next _lowerCamelCase : List[str] = new_node if index == len(self ) - 1: # insert at tail _lowerCamelCase : Any = new_node def A_ ( self ): return self.delete_nth(0 ) def A_ ( self ): return self.delete_nth(len(self ) - 1 ) def A_ ( self , lowercase = 0 ): if not 0 <= index < len(self ): raise IndexError('list index out of range.' ) _lowerCamelCase : Any = self.head if self.head == self.tail: # just one node _lowerCamelCase : List[str] = None elif index == 0: # delete head node _lowerCamelCase : List[str] = self.tail.next.next _lowerCamelCase : Optional[int] = self.head.next else: _lowerCamelCase : Dict = self.head for _ in range(index - 1 ): _lowerCamelCase : List[Any] = temp.next _lowerCamelCase : int = temp.next _lowerCamelCase : Optional[int] = temp.next.next if index == len(self ) - 1: # delete at tail _lowerCamelCase : List[Any] = temp return delete_node.data def A_ ( self ): return len(self ) == 0 def _snake_case ( ): _lowerCamelCase : Union[str, Any] = CircularLinkedList() assert len(lowercase__ ) == 0 assert circular_linked_list.is_empty() is True assert str(lowercase__ ) == "" try: circular_linked_list.delete_front() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_tail() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_nth(-1 ) raise AssertionError except IndexError: assert True try: circular_linked_list.delete_nth(0 ) raise AssertionError except IndexError: assert True assert circular_linked_list.is_empty() is True for i in range(5 ): assert len(lowercase__ ) == i circular_linked_list.insert_nth(lowercase__ , i + 1 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 6 ) ) circular_linked_list.insert_tail(6 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 7 ) ) circular_linked_list.insert_head(0 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(0 , 7 ) ) assert circular_linked_list.delete_front() == 0 assert circular_linked_list.delete_tail() == 6 assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 6 ) ) assert circular_linked_list.delete_nth(2 ) == 3 circular_linked_list.insert_nth(2 , 3 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 6 ) ) assert circular_linked_list.is_empty() is False if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" import argparse import os import torch from transformers.utils import WEIGHTS_NAME lowercase__ = ["""small""", """medium""", """large"""] lowercase__ = """lm_head.decoder.weight""" lowercase__ = """lm_head.weight""" def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : Any = torch.load(lowercase__ ) _lowerCamelCase : List[Any] = d.pop(lowercase__ ) os.makedirs(lowercase__ , exist_ok=lowercase__ ) torch.save(lowercase__ , os.path.join(lowercase__ , lowercase__ ) ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() parser.add_argument("""--dialogpt_path""", default=""".""", type=str) lowercase__ = parser.parse_args() for MODEL in DIALOGPT_MODELS: lowercase__ = os.path.join(args.dialogpt_path, F"{MODEL}_ft.pkl") lowercase__ = F"./DialoGPT-{MODEL}" convert_dialogpt_checkpoint( checkpoint_path, pytorch_dump_folder_path, )
96
"""simple docstring""" # Lint as: python3 import os import re import urllib.parse from pathlib import Path from typing import Callable, List, Optional, Union from zipfile import ZipFile from ..utils.file_utils import cached_path, hf_github_url from ..utils.logging import get_logger from ..utils.version import Version lowercase__ = get_logger(__name__) class lowerCAmelCase__ : '''simple docstring''' lowerCamelCase__ = """dummy_data""" lowerCamelCase__ = """datasets""" lowerCamelCase__ = False def __init__( self , lowercase , lowercase , lowercase , lowercase = None , lowercase = False , lowercase = True , lowercase = None , ): _lowerCamelCase : Optional[Any] = 0 _lowerCamelCase : Dict = dataset_name _lowerCamelCase : Union[str, Any] = cache_dir _lowerCamelCase : Dict = use_local_dummy_data _lowerCamelCase : Tuple = config # download_callbacks take a single url as input _lowerCamelCase : List[Callable] = download_callbacks or [] # if False, it doesn't load existing files and it returns the paths of the dummy files relative # to the dummy_data zip file root _lowerCamelCase : Any = load_existing_dummy_data # TODO(PVP, QL) might need to make this more general _lowerCamelCase : str = str(lowercase ) # to be downloaded _lowerCamelCase : Union[str, Any] = None _lowerCamelCase : int = None @property def A_ ( self ): if self._dummy_file is None: _lowerCamelCase : Tuple = self.download_dummy_data() return self._dummy_file @property def A_ ( self ): if self.config is not None: # structure is dummy / config_name / version_name return os.path.join('dummy' , self.config.name , self.version_name ) # structure is dummy / version_name return os.path.join('dummy' , self.version_name ) @property def A_ ( self ): return os.path.join(self.dummy_data_folder , 'dummy_data.zip' ) def A_ ( self ): _lowerCamelCase : List[str] = ( self.local_path_to_dummy_data if self.use_local_dummy_data is True else self.github_path_to_dummy_data ) _lowerCamelCase : int = cached_path( lowercase , cache_dir=self.cache_dir , extract_compressed_file=lowercase , force_extract=lowercase ) return os.path.join(lowercase , self.dummy_file_name ) @property def A_ ( self ): return os.path.join(self.datasets_scripts_dir , self.dataset_name , self.dummy_zip_file ) @property def A_ ( self ): if self._bucket_url is None: _lowerCamelCase : List[Any] = hf_github_url(self.dataset_name , self.dummy_zip_file.replace(os.sep , '/' ) ) return self._bucket_url @property def A_ ( self ): # return full path if its a dir if os.path.isdir(self.dummy_file ): return self.dummy_file # else cut off path to file -> example `xsum`. return "/".join(self.dummy_file.replace(os.sep , '/' ).split('/' )[:-1] ) def A_ ( self , lowercase , *lowercase ): if self.load_existing_dummy_data: # dummy data is downloaded and tested _lowerCamelCase : Union[str, Any] = self.dummy_file else: # dummy data cannot be downloaded and only the path to dummy file is returned _lowerCamelCase : Union[str, Any] = self.dummy_file_name # special case when data_url is a dict if isinstance(lowercase , lowercase ): return self.create_dummy_data_dict(lowercase , lowercase ) elif isinstance(lowercase , (list, tuple) ): return self.create_dummy_data_list(lowercase , lowercase ) else: return self.create_dummy_data_single(lowercase , lowercase ) def A_ ( self , lowercase , *lowercase ): return self.download_and_extract(lowercase ) def A_ ( self , lowercase , lowercase ): return self.download_and_extract(lowercase ) def A_ ( self , lowercase , *lowercase , **lowercase ): return path def A_ ( self ): return {} def A_ ( self , lowercase , lowercase ): _lowerCamelCase : Optional[int] = {} for key, single_urls in data_url.items(): for download_callback in self.download_callbacks: if isinstance(lowercase , lowercase ): for single_url in single_urls: download_callback(lowercase ) else: _lowerCamelCase : List[Any] = single_urls download_callback(lowercase ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus if isinstance(lowercase , lowercase ): _lowerCamelCase : List[Any] = [os.path.join(lowercase , urllib.parse.quote_plus(Path(lowercase ).name ) ) for x in single_urls] else: _lowerCamelCase : Optional[int] = single_urls _lowerCamelCase : List[Any] = os.path.join(lowercase , urllib.parse.quote_plus(Path(lowercase ).name ) ) _lowerCamelCase : int = value # make sure that values are unique if all(isinstance(lowercase , lowercase ) for i in dummy_data_dict.values() ) and len(set(dummy_data_dict.values() ) ) < len( dummy_data_dict.values() ): # append key to value to make its name unique _lowerCamelCase : List[Any] = {key: value + key for key, value in dummy_data_dict.items()} return dummy_data_dict def A_ ( self , lowercase , lowercase ): _lowerCamelCase : Optional[Any] = [] # trick: if there are many shards named like `data.txt-000001-of-00300`, only use the first one _lowerCamelCase : List[str] = all(bool(re.findall('[0-9]{3,}-of-[0-9]{3,}' , lowercase ) ) for url in data_url ) _lowerCamelCase : int = all( url.startswith('https://ftp.ncbi.nlm.nih.gov/pubmed/baseline/pubmed' ) for url in data_url ) if data_url and (is_tf_records or is_pubmed_records): _lowerCamelCase : List[str] = [data_url[0]] * len(lowercase ) for single_url in data_url: for download_callback in self.download_callbacks: download_callback(lowercase ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus _lowerCamelCase : str = os.path.join(lowercase , urllib.parse.quote_plus(single_url.split('/' )[-1] ) ) dummy_data_list.append(lowercase ) return dummy_data_list def A_ ( self , lowercase , lowercase ): for download_callback in self.download_callbacks: download_callback(lowercase ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus _lowerCamelCase : Tuple = os.path.join(lowercase , urllib.parse.quote_plus(data_url.split('/' )[-1] ) ) if os.path.exists(lowercase ) or not self.load_existing_dummy_data: return value else: # Backward compatibility, maybe deprecate at one point. # For many datasets with single url calls to dl_manager.download_and_extract, # the dummy_data.zip file is actually the zipped downloaded file # while now we expected the dummy_data.zip file to be a directory containing # the downloaded file. return path_to_dummy_data def A_ ( self ): pass def A_ ( self ): pass def A_ ( self , lowercase ): def _iter_archive_members(lowercase ): # this preserves the order of the members inside the ZIP archive _lowerCamelCase : str = Path(self.dummy_file ).parent _lowerCamelCase : Union[str, Any] = path.relative_to(lowercase ) with ZipFile(self.local_path_to_dummy_data ) as zip_file: _lowerCamelCase : List[str] = zip_file.namelist() for member in members: if member.startswith(relative_path.as_posix() ): yield dummy_parent_path.joinpath(lowercase ) _lowerCamelCase : Optional[int] = Path(lowercase ) _lowerCamelCase : Dict = _iter_archive_members(lowercase ) if self.use_local_dummy_data else path.rglob('*' ) for file_path in file_paths: if file_path.is_file() and not file_path.name.startswith(('.', '__') ): yield file_path.relative_to(lowercase ).as_posix(), file_path.open('rb' ) def A_ ( self , lowercase ): if not isinstance(lowercase , lowercase ): _lowerCamelCase : List[str] = [paths] for path in paths: if os.path.isfile(lowercase ): if os.path.basename(lowercase ).startswith(('.', '__') ): return yield path else: for dirpath, dirnames, filenames in os.walk(lowercase ): if os.path.basename(lowercase ).startswith(('.', '__') ): continue dirnames.sort() for filename in sorted(lowercase ): if filename.startswith(('.', '__') ): continue yield os.path.join(lowercase , lowercase )
96
1
"""simple docstring""" # Copyright (c) 2021-, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. #################################################################################################### # # Note: If when running this conversion script you're getting an exception: # ModuleNotFoundError: No module named 'megatron.model.enums' # you need to tell python where to find the clone of Megatron-LM, e.g.: # # cd /tmp # git clone https://github.com/NVIDIA/Megatron-LM # PYTHONPATH=/tmp/Megatron-LM python src/transformers/models/megatron_gpt2/convert_megatron_gpt2_checkpoint.py ... # # if you already have it cloned elsewhere, simply adjust the path to the existing path # # If the training was done using a Megatron-LM fork, e.g., # https://github.com/microsoft/Megatron-DeepSpeed/ then chances are that you need to have that one # in your path, i.e., /path/to/Megatron-DeepSpeed/ # import argparse import os import re import zipfile import torch from transformers import AutoTokenizer, GPTaConfig def _snake_case ( lowercase__ , lowercase__ , lowercase__=0 ): # Format the message. if name is None: _lowerCamelCase : Optional[Any] = None else: _lowerCamelCase : List[Any] = '.' * max(0 , spaces - 2 ) + '# {:' + str(50 - spaces ) + 's}' _lowerCamelCase : Optional[Any] = fmt.format(lowercase__ ) # Print and recurse (if needed). if isinstance(lowercase__ , lowercase__ ): if msg is not None: print(lowercase__ ) for k in val.keys(): recursive_print(lowercase__ , val[k] , spaces + 2 ) elif isinstance(lowercase__ , torch.Tensor ): print(lowercase__ , ':' , val.size() ) else: print(lowercase__ , ':' , lowercase__ ) def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): # Permutes layout of param tensor to [num_splits * num_heads * hidden_size, :] # for compatibility with later versions of NVIDIA Megatron-LM. # The inverse operation is performed inside Megatron-LM to read checkpoints: # https://github.com/NVIDIA/Megatron-LM/blob/v2.4/megatron/checkpointing.py#L209 # If param is the weight tensor of the self-attention block, the returned tensor # will have to be transposed one more time to be read by HuggingFace GPT2. _lowerCamelCase : Optional[Any] = param.size() if checkpoint_version == 1.0: # version 1.0 stores [num_heads * hidden_size * num_splits, :] _lowerCamelCase : List[Any] = (num_heads, hidden_size, num_splits) + input_shape[1:] _lowerCamelCase : List[Any] = param.view(*lowercase__ ) _lowerCamelCase : List[Any] = param.transpose(0 , 2 ) _lowerCamelCase : int = param.transpose(1 , 2 ).contiguous() elif checkpoint_version >= 2.0: # other versions store [num_heads * num_splits * hidden_size, :] _lowerCamelCase : List[Any] = (num_heads, num_splits, hidden_size) + input_shape[1:] _lowerCamelCase : Dict = param.view(*lowercase__ ) _lowerCamelCase : Union[str, Any] = param.transpose(0 , 1 ).contiguous() _lowerCamelCase : Any = param.view(*lowercase__ ) return param def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): # The converted output model. _lowerCamelCase : List[Any] = {} # old versions did not store training args _lowerCamelCase : List[Any] = input_state_dict.get('args' , lowercase__ ) if ds_args is not None: # do not make the user write a config file when the exact dimensions/sizes are already in the checkpoint # from pprint import pprint # pprint(vars(ds_args)) _lowerCamelCase : Union[str, Any] = ds_args.padded_vocab_size _lowerCamelCase : Dict = ds_args.max_position_embeddings _lowerCamelCase : Optional[int] = ds_args.hidden_size _lowerCamelCase : Optional[int] = ds_args.num_layers _lowerCamelCase : Optional[int] = ds_args.num_attention_heads _lowerCamelCase : Tuple = ds_args.ffn_hidden_size # pprint(config) # The number of heads. _lowerCamelCase : int = config.n_head # The hidden_size per head. _lowerCamelCase : Tuple = config.n_embd // config.n_head # Megatron-LM checkpoint version if "checkpoint_version" in input_state_dict.keys(): _lowerCamelCase : Union[str, Any] = input_state_dict['checkpoint_version'] else: _lowerCamelCase : List[Any] = 0.0 # The model. _lowerCamelCase : Optional[Any] = input_state_dict['model'] # The language model. _lowerCamelCase : Optional[Any] = model['language_model'] # The embeddings. _lowerCamelCase : Optional[Any] = lm['embedding'] # The word embeddings. _lowerCamelCase : Dict = embeddings['word_embeddings']['weight'] # Truncate the embedding table to vocab_size rows. _lowerCamelCase : Optional[int] = word_embeddings[: config.vocab_size, :] _lowerCamelCase : Dict = word_embeddings # The position embeddings. _lowerCamelCase : int = embeddings['position_embeddings']['weight'] # Read the causal mask dimension (seqlen). [max_sequence_length, hidden_size] _lowerCamelCase : Union[str, Any] = pos_embeddings.size(0 ) if n_positions != config.n_positions: raise ValueError( f'''pos_embeddings.max_sequence_length={n_positions} and config.n_positions={config.n_positions} don\'t match''' ) # Store the position embeddings. _lowerCamelCase : Optional[Any] = pos_embeddings # The transformer. _lowerCamelCase : Optional[Any] = lm['transformer'] if 'transformer' in lm.keys() else lm['encoder'] # The regex to extract layer names. _lowerCamelCase : Optional[Any] = re.compile(r'layers\.(\d+)\.([a-z0-9_.]+)\.([a-z]+)' ) # The simple map of names for "automated" rules. _lowerCamelCase : Any = { 'attention.dense': '.attn.c_proj.', 'self_attention.dense': '.attn.c_proj.', 'mlp.dense_h_to_4h': '.mlp.c_fc.', 'mlp.dense_4h_to_h': '.mlp.c_proj.', } # Extract the layers. for key, val in transformer.items(): # Match the name. _lowerCamelCase : List[str] = layer_re.match(lowercase__ ) # Stop if that's not a layer if m is None: break # The index of the layer. _lowerCamelCase : Union[str, Any] = int(m.group(1 ) ) # The name of the operation. _lowerCamelCase : str = m.group(2 ) # Is it a weight or a bias? _lowerCamelCase : List[Any] = m.group(3 ) # The name of the layer. _lowerCamelCase : Optional[Any] = f'''transformer.h.{layer_idx}''' # For layernorm(s), simply store the layer norm. if op_name.endswith('layernorm' ): _lowerCamelCase : Dict = 'ln_1' if op_name.startswith('input' ) else 'ln_2' _lowerCamelCase : List[Any] = val # Transpose the QKV matrix. elif ( op_name == "attention.query_key_value" or op_name == "self_attention.query_key_value" ) and weight_or_bias == "weight": # Insert a tensor of 1x1xDxD bias. _lowerCamelCase : Optional[int] = torch.tril(torch.ones((n_positions, n_positions) , dtype=torch.floataa ) ).view( 1 , 1 , lowercase__ , lowercase__ ) _lowerCamelCase : Optional[Any] = causal_mask # Insert a "dummy" tensor for masked_bias. _lowerCamelCase : Dict = torch.tensor(-1E4 , dtype=torch.floataa ) _lowerCamelCase : Tuple = masked_bias _lowerCamelCase : Dict = fix_query_key_value_ordering(lowercase__ , lowercase__ , 3 , lowercase__ , lowercase__ ) # Megatron stores (3*D) x D but transformers-GPT2 expects D x 3*D. _lowerCamelCase : Dict = out_val.transpose(0 , 1 ).contiguous() # Store. _lowerCamelCase : int = out_val # Transpose the bias. elif ( op_name == "attention.query_key_value" or op_name == "self_attention.query_key_value" ) and weight_or_bias == "bias": _lowerCamelCase : Dict = fix_query_key_value_ordering(lowercase__ , lowercase__ , 3 , lowercase__ , lowercase__ ) # Store. No change of shape. _lowerCamelCase : List[str] = out_val # Transpose the weights. elif weight_or_bias == "weight": _lowerCamelCase : Optional[int] = megatron_to_transformers[op_name] _lowerCamelCase : List[Any] = val.transpose(0 , 1 ) # Copy the bias. elif weight_or_bias == "bias": _lowerCamelCase : Dict = megatron_to_transformers[op_name] _lowerCamelCase : Union[str, Any] = val # DEBUG. assert config.n_layer == layer_idx + 1 # The final layernorm. _lowerCamelCase : Any = transformer['final_layernorm.weight'] _lowerCamelCase : Union[str, Any] = transformer['final_layernorm.bias'] # For LM head, transformers' wants the matrix to weight embeddings. _lowerCamelCase : int = word_embeddings # It should be done! return output_state_dict def _snake_case ( ): # Create the argument parser. _lowerCamelCase : Union[str, Any] = argparse.ArgumentParser() parser.add_argument('--print-checkpoint-structure' , action='store_true' ) parser.add_argument( 'path_to_checkpoint' , type=lowercase__ , help='Path to the checkpoint file (.zip archive or direct .pt file)' , ) parser.add_argument( '--config_file' , default='' , type=lowercase__ , help='An optional config json file describing the pre-trained model.' , ) _lowerCamelCase : Dict = parser.parse_args() # Extract the basename. _lowerCamelCase : Optional[int] = os.path.dirname(args.path_to_checkpoint ) # Load the model. # the .zip is very optional, let's keep it for backward compatibility print(f'''Extracting PyTorch state dictionary from {args.path_to_checkpoint}''' ) if args.path_to_checkpoint.endswith('.zip' ): with zipfile.ZipFile(args.path_to_checkpoint , 'r' ) as checkpoint: with checkpoint.open('release/mp_rank_00/model_optim_rng.pt' ) as pytorch_dict: _lowerCamelCase : Any = torch.load(lowercase__ , map_location='cpu' ) else: _lowerCamelCase : int = torch.load(args.path_to_checkpoint , map_location='cpu' ) _lowerCamelCase : List[Any] = input_state_dict.get('args' , lowercase__ ) # Read the config, or default to the model released by NVIDIA. if args.config_file == "": if ds_args is not None: if ds_args.bias_gelu_fusion: _lowerCamelCase : Union[str, Any] = 'gelu_fast' elif ds_args.openai_gelu: _lowerCamelCase : Tuple = 'gelu_new' else: _lowerCamelCase : int = 'gelu' else: # in the very early days this used to be "gelu_new" _lowerCamelCase : int = 'gelu_new' # Spell out all parameters in case the defaults change. _lowerCamelCase : Tuple = GPTaConfig( vocab_size=50257 , n_positions=1024 , n_embd=1024 , n_layer=24 , n_head=16 , n_inner=4096 , activation_function=lowercase__ , resid_pdrop=0.1 , embd_pdrop=0.1 , attn_pdrop=0.1 , layer_norm_epsilon=1E-5 , initializer_range=0.0_2 , summary_type='cls_index' , summary_use_proj=lowercase__ , summary_activation=lowercase__ , summary_proj_to_labels=lowercase__ , summary_first_dropout=0.1 , scale_attn_weights=lowercase__ , use_cache=lowercase__ , bos_token_id=50256 , eos_token_id=50256 , ) else: _lowerCamelCase : Union[str, Any] = GPTaConfig.from_json_file(args.config_file ) _lowerCamelCase : Any = ['GPT2LMHeadModel'] # Convert. print('Converting' ) _lowerCamelCase : int = convert_megatron_checkpoint(lowercase__ , lowercase__ , lowercase__ ) # Print the structure of converted state dict. if args.print_checkpoint_structure: recursive_print(lowercase__ , lowercase__ ) # Add tokenizer class info to config # see https://github.com/huggingface/transformers/issues/13906) if ds_args is not None: _lowerCamelCase : Optional[int] = ds_args.tokenizer_type if tokenizer_type == "GPT2BPETokenizer": _lowerCamelCase : Optional[int] = 'gpt2' elif tokenizer_type == "PretrainedFromHF": _lowerCamelCase : Any = ds_args.tokenizer_name_or_path else: raise ValueError(f'''Unrecognized tokenizer_type {tokenizer_type}''' ) else: _lowerCamelCase : Any = 'gpt2' _lowerCamelCase : Optional[Any] = AutoTokenizer.from_pretrained(lowercase__ ) _lowerCamelCase : List[Any] = type(lowercase__ ).__name__ _lowerCamelCase : Tuple = tokenizer_class # Store the config to file. print('Saving config' ) config.save_pretrained(lowercase__ ) # Save tokenizer based on args print(f'''Adding {tokenizer_class} tokenizer files''' ) tokenizer.save_pretrained(lowercase__ ) # Store the state_dict to file. _lowerCamelCase : Dict = os.path.join(lowercase__ , 'pytorch_model.bin' ) print(f'''Saving checkpoint to "{output_checkpoint_file}"''' ) torch.save(lowercase__ , lowercase__ ) #################################################################################################### if __name__ == "__main__": main() ####################################################################################################
96
"""simple docstring""" def _snake_case ( lowercase__ ): stooge(lowercase__ , 0 , len(lowercase__ ) - 1 ) return arr def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: _lowerCamelCase, _lowerCamelCase : Optional[Any] = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: _lowerCamelCase : Union[str, Any] = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(lowercase__ , lowercase__ , (h - t) ) # Recursively sort last 2/3 elements stooge(lowercase__ , i + t , (lowercase__) ) # Recursively sort first 2/3 elements stooge(lowercase__ , lowercase__ , (h - t) ) if __name__ == "__main__": lowercase__ = input("""Enter numbers separated by a comma:\n""").strip() lowercase__ = [int(item) for item in user_input.split(""",""")] print(stooge_sort(unsorted))
96
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_deit import DeiTImageProcessor lowercase__ = logging.get_logger(__name__) class lowerCAmelCase__ ( lowercase ): '''simple docstring''' def __init__( self , *lowercase , **lowercase ): warnings.warn( 'The class DeiTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use DeiTImageProcessor instead.' , lowercase , ) super().__init__(*lowercase , **lowercase )
96
"""simple docstring""" import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = ["""image_processor""", """tokenizer"""] lowerCamelCase__ = """BlipImageProcessor""" lowerCamelCase__ = """AutoTokenizer""" def __init__( self , lowercase , lowercase , lowercase ): super().__init__(lowercase , lowercase ) # add QFormer tokenizer _lowerCamelCase : int = qformer_tokenizer def __call__( self , lowercase = None , lowercase = None , lowercase = True , lowercase = False , lowercase = None , lowercase = None , lowercase = 0 , lowercase = None , lowercase = None , lowercase = False , lowercase = False , lowercase = False , lowercase = False , lowercase = False , lowercase = True , lowercase = None , **lowercase , ): if images is None and text is None: raise ValueError('You have to specify at least images or text.' ) _lowerCamelCase : int = BatchFeature() if text is not None: _lowerCamelCase : List[str] = self.tokenizer( text=lowercase , add_special_tokens=lowercase , padding=lowercase , truncation=lowercase , max_length=lowercase , stride=lowercase , pad_to_multiple_of=lowercase , return_attention_mask=lowercase , return_overflowing_tokens=lowercase , return_special_tokens_mask=lowercase , return_offsets_mapping=lowercase , return_token_type_ids=lowercase , return_length=lowercase , verbose=lowercase , return_tensors=lowercase , **lowercase , ) encoding.update(lowercase ) _lowerCamelCase : List[str] = self.qformer_tokenizer( text=lowercase , add_special_tokens=lowercase , padding=lowercase , truncation=lowercase , max_length=lowercase , stride=lowercase , pad_to_multiple_of=lowercase , return_attention_mask=lowercase , return_overflowing_tokens=lowercase , return_special_tokens_mask=lowercase , return_offsets_mapping=lowercase , return_token_type_ids=lowercase , return_length=lowercase , verbose=lowercase , return_tensors=lowercase , **lowercase , ) _lowerCamelCase : List[Any] = qformer_text_encoding.pop('input_ids' ) _lowerCamelCase : Tuple = qformer_text_encoding.pop('attention_mask' ) if images is not None: _lowerCamelCase : int = self.image_processor(lowercase , return_tensors=lowercase ) encoding.update(lowercase ) return encoding def A_ ( self , *lowercase , **lowercase ): return self.tokenizer.batch_decode(*lowercase , **lowercase ) def A_ ( self , *lowercase , **lowercase ): return self.tokenizer.decode(*lowercase , **lowercase ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def A_ ( self ): _lowerCamelCase : Union[str, Any] = self.tokenizer.model_input_names _lowerCamelCase : Any = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def A_ ( self , lowercase , **lowercase ): if os.path.isfile(lowercase ): raise ValueError(F'''Provided path ({save_directory}) should be a directory, not a file''' ) os.makedirs(lowercase , exist_ok=lowercase ) _lowerCamelCase : Optional[Any] = os.path.join(lowercase , 'qformer_tokenizer' ) self.qformer_tokenizer.save_pretrained(lowercase ) return super().save_pretrained(lowercase , **lowercase ) @classmethod def A_ ( cls , lowercase , **lowercase ): _lowerCamelCase : Union[str, Any] = AutoTokenizer.from_pretrained(lowercase , subfolder='qformer_tokenizer' ) _lowerCamelCase : Dict = cls._get_arguments_from_pretrained(lowercase , **lowercase ) args.append(lowercase ) return cls(*lowercase )
96
1
"""simple docstring""" import gc import threading import time import psutil import torch class lowerCAmelCase__ : '''simple docstring''' def __init__( self ): _lowerCamelCase : Tuple = psutil.Process() _lowerCamelCase : Optional[int] = False def A_ ( self ): _lowerCamelCase : Tuple = -1 while True: _lowerCamelCase : List[Any] = max(self.process.memory_info().rss , self.cpu_memory_peak ) # can't sleep or will not catch the peak right (this comment is here on purpose) if not self.peak_monitoring: break def A_ ( self ): _lowerCamelCase : List[str] = True _lowerCamelCase : Union[str, Any] = threading.Thread(target=self.peak_monitor ) _lowerCamelCase : Optional[Any] = True self.thread.start() def A_ ( self ): _lowerCamelCase : Optional[Any] = False self.thread.join() return self.cpu_memory_peak lowercase__ = PeakCPUMemory() def _snake_case ( ): # Time _lowerCamelCase : Dict = {'time': time.time()} gc.collect() torch.cuda.empty_cache() # CPU mem _lowerCamelCase : str = psutil.Process().memory_info().rss cpu_peak_tracker.start() # GPU mem for i in range(torch.cuda.device_count() ): _lowerCamelCase : List[Any] = torch.cuda.memory_allocated(lowercase__ ) torch.cuda.reset_peak_memory_stats() return measures def _snake_case ( lowercase__ ): # Time _lowerCamelCase : Any = {'time': time.time() - start_measures['time']} gc.collect() torch.cuda.empty_cache() # CPU mem _lowerCamelCase : Dict = (psutil.Process().memory_info().rss - start_measures['cpu']) / 2**20 _lowerCamelCase : str = (cpu_peak_tracker.stop() - start_measures['cpu']) / 2**20 # GPU mem for i in range(torch.cuda.device_count() ): _lowerCamelCase : List[Any] = (torch.cuda.memory_allocated(lowercase__ ) - start_measures[str(lowercase__ )]) / 2**20 _lowerCamelCase : Union[str, Any] = (torch.cuda.max_memory_allocated(lowercase__ ) - start_measures[str(lowercase__ )]) / 2**20 return measures def _snake_case ( lowercase__ , lowercase__ ): print(f'''{description}:''' ) print(f'''- Time: {measures['time']:.2f}s''' ) for i in range(torch.cuda.device_count() ): print(f'''- GPU {i} allocated: {measures[str(lowercase__ )]:.2f}MiB''' ) _lowerCamelCase : Any = measures[f'''{i}-peak'''] print(f'''- GPU {i} peak: {peak:.2f}MiB''' ) print(f'''- CPU RAM allocated: {measures['cpu']:.2f}MiB''' ) print(f'''- CPU RAM peak: {measures['cpu-peak']:.2f}MiB''' )
96
"""simple docstring""" import json import logging import os import sys from time import time from unittest.mock import patch from transformers.testing_utils import TestCasePlus, require_torch_tpu logging.basicConfig(level=logging.DEBUG) lowercase__ = logging.getLogger() def _snake_case ( lowercase__ ): _lowerCamelCase : List[Any] = {} _lowerCamelCase : List[Any] = os.path.join(lowercase__ , 'all_results.json' ) if os.path.exists(lowercase__ ): with open(lowercase__ , 'r' ) as f: _lowerCamelCase : List[Any] = json.load(lowercase__ ) else: raise ValueError(f'''can\'t find {path}''' ) return results lowercase__ = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) @require_torch_tpu class lowerCAmelCase__ ( lowercase ): '''simple docstring''' def A_ ( self ): import xla_spawn _lowerCamelCase : List[Any] = self.get_auto_remove_tmp_dir() _lowerCamelCase : List[Any] = F''' ./examples/pytorch/text-classification/run_glue.py --num_cores=8 ./examples/pytorch/text-classification/run_glue.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --overwrite_output_dir --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --do_train --do_eval --debug tpu_metrics_debug --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --max_steps=10 --warmup_steps=2 --seed=42 --max_seq_length=128 '''.split() with patch.object(lowercase , 'argv' , lowercase ): _lowerCamelCase : Dict = time() xla_spawn.main() _lowerCamelCase : Any = time() _lowerCamelCase : Optional[int] = get_results(lowercase ) self.assertGreaterEqual(result['eval_accuracy'] , 0.75 ) # Assert that the script takes less than 500 seconds to make sure it doesn't hang. self.assertLess(end - start , 500 ) def A_ ( self ): import xla_spawn _lowerCamelCase : Tuple = '\n ./tests/test_trainer_tpu.py\n --num_cores=8\n ./tests/test_trainer_tpu.py\n '.split() with patch.object(lowercase , 'argv' , lowercase ): xla_spawn.main()
96
1
"""simple docstring""" from typing import Callable, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { """microsoft/xprophetnet-large-wiki100-cased""": ( """https://huggingface.co/microsoft/xprophetnet-large-wiki100-cased/resolve/main/config.json""" ), } class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """xlm-prophetnet""" lowerCamelCase__ = ["""past_key_values"""] lowerCamelCase__ = { """num_attention_heads""": """num_encoder_attention_heads""", } def __init__( self , lowercase = 0.1 , lowercase = "gelu" , lowercase = 30522 , lowercase = 1024 , lowercase = 4096 , lowercase = 12 , lowercase = 16 , lowercase = 4096 , lowercase = 12 , lowercase = 16 , lowercase = 0.1 , lowercase = 0.1 , lowercase = 512 , lowercase = 0.02 , lowercase = True , lowercase = True , lowercase = 0 , lowercase = 2 , lowercase = 32 , lowercase = 128 , lowercase = False , lowercase = 0.0 , lowercase = True , lowercase = 0 , lowercase = 1 , lowercase = 2 , **lowercase , ): _lowerCamelCase : Tuple = vocab_size _lowerCamelCase : Dict = hidden_size _lowerCamelCase : List[str] = encoder_ffn_dim _lowerCamelCase : Optional[int] = num_encoder_layers _lowerCamelCase : Tuple = num_encoder_attention_heads _lowerCamelCase : Dict = decoder_ffn_dim _lowerCamelCase : Optional[Any] = num_decoder_layers _lowerCamelCase : Any = num_decoder_attention_heads _lowerCamelCase : Optional[int] = max_position_embeddings _lowerCamelCase : int = init_std # Normal(0, this parameter) _lowerCamelCase : int = activation_function # parameters for xlmprophetnet _lowerCamelCase : List[str] = ngram _lowerCamelCase : Tuple = num_buckets _lowerCamelCase : int = relative_max_distance _lowerCamelCase : str = disable_ngram_loss _lowerCamelCase : Union[str, Any] = eps # 3 Types of Dropout _lowerCamelCase : str = attention_dropout _lowerCamelCase : List[Any] = activation_dropout _lowerCamelCase : Any = dropout _lowerCamelCase : List[Any] = use_cache super().__init__( pad_token_id=lowercase , bos_token_id=lowercase , eos_token_id=lowercase , is_encoder_decoder=lowercase , add_cross_attention=lowercase , decoder_start_token_id=lowercase , **lowercase , ) @property def A_ ( self ): return self.num_encoder_layers + self.num_decoder_layers @num_hidden_layers.setter def A_ ( self , lowercase ): raise NotImplementedError( 'This model does not support the setting of `num_hidden_layers`. Please set `num_encoder_layers` and' ' `num_decoder_layers`.' )
96
"""simple docstring""" from __future__ import annotations import math import numpy as np from numpy.linalg import norm def _snake_case ( lowercase__ , lowercase__ ): return math.sqrt(sum(pow(a - b , 2 ) for a, b in zip(lowercase__ , lowercase__ ) ) ) def _snake_case ( lowercase__ , lowercase__ ): if dataset.ndim != value_array.ndim: _lowerCamelCase : Tuple = ( 'Wrong input data\'s dimensions... ' f'''dataset : {dataset.ndim}, value_array : {value_array.ndim}''' ) raise ValueError(lowercase__ ) try: if dataset.shape[1] != value_array.shape[1]: _lowerCamelCase : Optional[int] = ( 'Wrong input data\'s shape... ' f'''dataset : {dataset.shape[1]}, value_array : {value_array.shape[1]}''' ) raise ValueError(lowercase__ ) except IndexError: if dataset.ndim != value_array.ndim: raise TypeError('Wrong shape' ) if dataset.dtype != value_array.dtype: _lowerCamelCase : int = ( 'Input data have different datatype... ' f'''dataset : {dataset.dtype}, value_array : {value_array.dtype}''' ) raise TypeError(lowercase__ ) _lowerCamelCase : Optional[int] = [] for value in value_array: _lowerCamelCase : Tuple = euclidean(lowercase__ , dataset[0] ) _lowerCamelCase : Union[str, Any] = dataset[0].tolist() for dataset_value in dataset[1:]: _lowerCamelCase : Optional[Any] = euclidean(lowercase__ , lowercase__ ) if dist > temp_dist: _lowerCamelCase : List[Any] = temp_dist _lowerCamelCase : List[str] = dataset_value.tolist() answer.append([vector, dist] ) return answer def _snake_case ( lowercase__ , lowercase__ ): return np.dot(lowercase__ , lowercase__ ) / (norm(lowercase__ ) * norm(lowercase__ )) if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, CycleDiffusionPipeline, DDIMScheduler, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowerCAmelCase__ ( lowercase, lowercase, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = CycleDiffusionPipeline lowerCamelCase__ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { """negative_prompt""", """height""", """width""", """negative_prompt_embeds""", } lowerCamelCase__ = PipelineTesterMixin.required_optional_params - {"""latents"""} lowerCamelCase__ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"""source_prompt"""} ) lowerCamelCase__ = IMAGE_TO_IMAGE_IMAGE_PARAMS lowerCamelCase__ = IMAGE_TO_IMAGE_IMAGE_PARAMS def A_ ( self ): torch.manual_seed(0 ) _lowerCamelCase : Dict = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , ) _lowerCamelCase : Optional[Any] = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='scaled_linear' , num_train_timesteps=1000 , clip_sample=lowercase , set_alpha_to_one=lowercase , ) torch.manual_seed(0 ) _lowerCamelCase : Optional[Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) _lowerCamelCase : Optional[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) _lowerCamelCase : Any = CLIPTextModel(lowercase ) _lowerCamelCase : Optional[int] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) _lowerCamelCase : int = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def A_ ( self , lowercase , lowercase=0 ): _lowerCamelCase : Union[str, Any] = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowercase ) ).to(lowercase ) _lowerCamelCase : Optional[Any] = image / 2 + 0.5 if str(lowercase ).startswith('mps' ): _lowerCamelCase : Optional[int] = torch.manual_seed(lowercase ) else: _lowerCamelCase : Union[str, Any] = torch.Generator(device=lowercase ).manual_seed(lowercase ) _lowerCamelCase : List[Any] = { 'prompt': 'An astronaut riding an elephant', 'source_prompt': 'An astronaut riding a horse', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'eta': 0.1, 'strength': 0.8, 'guidance_scale': 3, 'source_guidance_scale': 1, 'output_type': 'numpy', } return inputs def A_ ( self ): _lowerCamelCase : Any = 'cpu' # ensure determinism for the device-dependent torch.Generator _lowerCamelCase : str = self.get_dummy_components() _lowerCamelCase : Tuple = CycleDiffusionPipeline(**lowercase ) _lowerCamelCase : int = pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : List[Any] = self.get_dummy_inputs(lowercase ) _lowerCamelCase : List[Any] = pipe(**lowercase ) _lowerCamelCase : Optional[int] = output.images _lowerCamelCase : Optional[Any] = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) _lowerCamelCase : List[str] = np.array([0.44_59, 0.49_43, 0.45_44, 0.66_43, 0.54_74, 0.43_27, 0.57_01, 0.59_59, 0.51_79] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @unittest.skipIf(torch_device != 'cuda' , 'This test requires a GPU' ) def A_ ( self ): _lowerCamelCase : Union[str, Any] = self.get_dummy_components() for name, module in components.items(): if hasattr(lowercase , 'half' ): _lowerCamelCase : str = module.half() _lowerCamelCase : List[str] = CycleDiffusionPipeline(**lowercase ) _lowerCamelCase : Union[str, Any] = pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : List[Any] = self.get_dummy_inputs(lowercase ) _lowerCamelCase : Union[str, Any] = pipe(**lowercase ) _lowerCamelCase : Tuple = output.images _lowerCamelCase : int = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) _lowerCamelCase : str = np.array([0.35_06, 0.45_43, 0.4_46, 0.45_75, 0.51_95, 0.41_55, 0.52_73, 0.5_18, 0.41_16] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @skip_mps def A_ ( self ): return super().test_save_load_local() @unittest.skip('non-deterministic pipeline' ) def A_ ( self ): return super().test_inference_batch_single_identical() @skip_mps def A_ ( self ): return super().test_dict_tuple_outputs_equivalent() @skip_mps def A_ ( self ): return super().test_save_load_optional_components() @skip_mps def A_ ( self ): return super().test_attention_slicing_forward_pass() @slow @require_torch_gpu class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def A_ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A_ ( self ): _lowerCamelCase : Union[str, Any] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/cycle-diffusion/black_colored_car.png' ) _lowerCamelCase : Any = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car_fp16.npy' ) _lowerCamelCase : Optional[Any] = init_image.resize((512, 512) ) _lowerCamelCase : Dict = 'CompVis/stable-diffusion-v1-4' _lowerCamelCase : Any = DDIMScheduler.from_pretrained(lowercase , subfolder='scheduler' ) _lowerCamelCase : Optional[int] = CycleDiffusionPipeline.from_pretrained( lowercase , scheduler=lowercase , safety_checker=lowercase , torch_dtype=torch.floataa , revision='fp16' ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) pipe.enable_attention_slicing() _lowerCamelCase : Tuple = 'A black colored car' _lowerCamelCase : Optional[Any] = 'A blue colored car' _lowerCamelCase : List[Any] = torch.manual_seed(0 ) _lowerCamelCase : str = pipe( prompt=lowercase , source_prompt=lowercase , image=lowercase , num_inference_steps=100 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=lowercase , output_type='np' , ) _lowerCamelCase : Optional[Any] = output.images # the values aren't exactly equal, but the images look the same visually assert np.abs(image - expected_image ).max() < 5E-1 def A_ ( self ): _lowerCamelCase : List[Any] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/cycle-diffusion/black_colored_car.png' ) _lowerCamelCase : str = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car.npy' ) _lowerCamelCase : Tuple = init_image.resize((512, 512) ) _lowerCamelCase : List[Any] = 'CompVis/stable-diffusion-v1-4' _lowerCamelCase : str = DDIMScheduler.from_pretrained(lowercase , subfolder='scheduler' ) _lowerCamelCase : str = CycleDiffusionPipeline.from_pretrained(lowercase , scheduler=lowercase , safety_checker=lowercase ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) pipe.enable_attention_slicing() _lowerCamelCase : List[str] = 'A black colored car' _lowerCamelCase : Tuple = 'A blue colored car' _lowerCamelCase : Dict = torch.manual_seed(0 ) _lowerCamelCase : Dict = pipe( prompt=lowercase , source_prompt=lowercase , image=lowercase , num_inference_steps=100 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=lowercase , output_type='np' , ) _lowerCamelCase : str = output.images assert np.abs(image - expected_image ).max() < 2E-2
96
"""simple docstring""" import socket def _snake_case ( ): _lowerCamelCase : List[Any] = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) _lowerCamelCase : Union[str, Any] = socket.gethostname() _lowerCamelCase : List[Any] = 12312 sock.connect((host, port) ) sock.send(B'Hello server!' ) with open('Received_file' , 'wb' ) as out_file: print('File opened' ) print('Receiving data...' ) while True: _lowerCamelCase : int = sock.recv(1024 ) if not data: break out_file.write(lowercase__ ) print('Successfully received the file' ) sock.close() print('Connection closed' ) if __name__ == "__main__": main()
96
1
"""simple docstring""" import argparse import os import re import packaging.version lowercase__ = """examples/""" lowercase__ = { """examples""": (re.compile(R"""^check_min_version\(\"[^\"]+\"\)\s*$""", re.MULTILINE), """check_min_version(\"VERSION\")\n"""), """init""": (re.compile(R"""^__version__\s+=\s+\"([^\"]+)\"\s*$""", re.MULTILINE), """__version__ = \"VERSION\"\n"""), """setup""": (re.compile(R"""^(\s*)version\s*=\s*\"[^\"]+\",""", re.MULTILINE), R"""\1version=\"VERSION\","""), """doc""": (re.compile(R"""^(\s*)release\s*=\s*\"[^\"]+\"$""", re.MULTILINE), """release = \"VERSION\"\n"""), } lowercase__ = { """init""": """src/transformers/__init__.py""", """setup""": """setup.py""", } lowercase__ = """README.md""" def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): with open(lowercase__ , 'r' , encoding='utf-8' , newline='\n' ) as f: _lowerCamelCase : Optional[int] = f.read() _lowerCamelCase, _lowerCamelCase : Optional[Any] = REPLACE_PATTERNS[pattern] _lowerCamelCase : Dict = replace.replace('VERSION' , lowercase__ ) _lowerCamelCase : int = re_pattern.sub(lowercase__ , lowercase__ ) with open(lowercase__ , 'w' , encoding='utf-8' , newline='\n' ) as f: f.write(lowercase__ ) def _snake_case ( lowercase__ ): for folder, directories, fnames in os.walk(lowercase__ ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('research_projects' ) if "legacy" in directories: directories.remove('legacy' ) for fname in fnames: if fname.endswith('.py' ): update_version_in_file(os.path.join(lowercase__ , lowercase__ ) , lowercase__ , pattern='examples' ) def _snake_case ( lowercase__ , lowercase__=False ): for pattern, fname in REPLACE_FILES.items(): update_version_in_file(lowercase__ , lowercase__ , lowercase__ ) if not patch: update_version_in_examples(lowercase__ ) def _snake_case ( ): _lowerCamelCase : int = '🤗 Transformers currently provides the following architectures' _lowerCamelCase : Tuple = '1. Want to contribute a new model?' with open(lowercase__ , 'r' , encoding='utf-8' , newline='\n' ) as f: _lowerCamelCase : int = f.readlines() # Find the start of the list. _lowerCamelCase : List[str] = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 _lowerCamelCase : List[str] = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith('1.' ): _lowerCamelCase : List[Any] = lines[index].replace( 'https://huggingface.co/docs/transformers/main/model_doc' , 'https://huggingface.co/docs/transformers/model_doc' , ) index += 1 with open(lowercase__ , 'w' , encoding='utf-8' , newline='\n' ) as f: f.writelines(lowercase__ ) def _snake_case ( ): with open(REPLACE_FILES['init'] , 'r' ) as f: _lowerCamelCase : Tuple = f.read() _lowerCamelCase : Union[str, Any] = REPLACE_PATTERNS['init'][0].search(lowercase__ ).groups()[0] return packaging.version.parse(lowercase__ ) def _snake_case ( lowercase__=False ): _lowerCamelCase : Optional[int] = get_version() if patch and default_version.is_devrelease: raise ValueError('Can\'t create a patch version from the dev branch, checkout a released version!' ) if default_version.is_devrelease: _lowerCamelCase : str = default_version.base_version elif patch: _lowerCamelCase : Union[str, Any] = f'''{default_version.major}.{default_version.minor}.{default_version.micro + 1}''' else: _lowerCamelCase : Any = f'''{default_version.major}.{default_version.minor + 1}.0''' # Now let's ask nicely if that's the right one. _lowerCamelCase : int = input(f'''Which version are you releasing? [{default_version}]''' ) if len(lowercase__ ) == 0: _lowerCamelCase : Tuple = default_version print(f'''Updating version to {version}.''' ) global_version_update(lowercase__ , patch=lowercase__ ) if not patch: print('Cleaning main README, don\'t forget to run `make fix-copies`.' ) clean_main_ref_in_model_list() def _snake_case ( ): _lowerCamelCase : Optional[Any] = get_version() _lowerCamelCase : Optional[int] = f'''{current_version.major}.{current_version.minor + 1}.0.dev0''' _lowerCamelCase : str = current_version.base_version # Check with the user we got that right. _lowerCamelCase : List[str] = input(f'''Which version are we developing now? [{dev_version}]''' ) if len(lowercase__ ) == 0: _lowerCamelCase : List[str] = dev_version print(f'''Updating version to {version}.''' ) global_version_update(lowercase__ ) print('Cleaning main README, don\'t forget to run `make fix-copies`.' ) clean_main_ref_in_model_list() if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() parser.add_argument("""--post_release""", action="""store_true""", help="""Whether this is pre or post release.""") parser.add_argument("""--patch""", action="""store_true""", help="""Whether or not this is a patch release.""") lowercase__ = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print("""Nothing to do after a patch :-)""") else: post_release_work()
96
"""simple docstring""" from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record lowercase__ = """\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } """ lowercase__ = """\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. """ lowercase__ = """ Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for 'record': list of question-answer dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'prediction_text': the predicted answer text - for 'multirc': list of question-answer dictionaries with the following keys: - 'idx': index of the question-answer pair as specified by the dataset - 'prediction': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for 'record': list of question-answers dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'answers': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for 'record': - 'exact_match': Exact match between answer and gold answer - 'f1': F1 score - for 'multirc': - 'exact_match': Exact match between answer and gold answer - 'f1_m': Per-question macro-F1 score - 'f1_a': Average F1 score over all answers - for 'axb': 'matthews_correlation': Matthew Correlation - for 'cb': - 'accuracy': Accuracy - 'f1': F1 score - for all others: - 'accuracy': Accuracy Examples: >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'cb') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'record') >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}] >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc') >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'axb') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def _snake_case ( lowercase__ , lowercase__ ): return float((preds == labels).mean() ) def _snake_case ( lowercase__ , lowercase__ , lowercase__="binary" ): _lowerCamelCase : str = simple_accuracy(lowercase__ , lowercase__ ) _lowerCamelCase : Any = float(fa_score(y_true=lowercase__ , y_pred=lowercase__ , average=lowercase__ ) ) return { "accuracy": acc, "f1": fa, } def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : Any = {} for id_pred, label in zip(lowercase__ , lowercase__ ): _lowerCamelCase : Tuple = f'''{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}''' _lowerCamelCase : Union[str, Any] = id_pred['prediction'] if question_id in question_map: question_map[question_id].append((pred, label) ) else: _lowerCamelCase : Optional[Any] = [(pred, label)] _lowerCamelCase, _lowerCamelCase : Optional[int] = [], [] for question, preds_labels in question_map.items(): _lowerCamelCase, _lowerCamelCase : Tuple = zip(*lowercase__ ) _lowerCamelCase : List[str] = fa_score(y_true=lowercase__ , y_pred=lowercase__ , average='macro' ) fas.append(lowercase__ ) _lowerCamelCase : int = int(sum(pred == label for pred, label in preds_labels ) == len(lowercase__ ) ) ems.append(lowercase__ ) _lowerCamelCase : Optional[Any] = float(sum(lowercase__ ) / len(lowercase__ ) ) _lowerCamelCase : Optional[int] = sum(lowercase__ ) / len(lowercase__ ) _lowerCamelCase : List[Any] = float(fa_score(y_true=lowercase__ , y_pred=[id_pred['prediction'] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCAmelCase__ ( datasets.Metric ): '''simple docstring''' def A_ ( self ): if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( 'You should supply a configuration name selected in ' '["boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg",]' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format='numpy' if not self.config_name == 'record' and not self.config_name == 'multirc' else None , ) def A_ ( self ): if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value('int64' ), "query": datasets.Value('int64' ), }, "prediction_text": datasets.Value('string' ), }, "references": { "idx": { "passage": datasets.Value('int64' ), "query": datasets.Value('int64' ), }, "answers": datasets.Sequence(datasets.Value('string' ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value('int64' ), "paragraph": datasets.Value('int64' ), "question": datasets.Value('int64' ), }, "prediction": datasets.Value('int64' ), }, "references": datasets.Value('int64' ), } else: return { "predictions": datasets.Value('int64' ), "references": datasets.Value('int64' ), } def A_ ( self , lowercase , lowercase ): if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(lowercase , lowercase )} elif self.config_name == "cb": return acc_and_fa(lowercase , lowercase , fa_avg='macro' ) elif self.config_name == "record": _lowerCamelCase : List[str] = [ { 'qas': [ {'id': ref['idx']['query'], 'answers': [{'text': ans} for ans in ref['answers']]} for ref in references ] } ] _lowerCamelCase : Union[str, Any] = {pred['idx']['query']: pred['prediction_text'] for pred in predictions} return evaluate_record(lowercase , lowercase )[0] elif self.config_name == "multirc": return evaluate_multirc(lowercase , lowercase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(lowercase , lowercase )} else: raise KeyError( 'You should supply a configuration name selected in ' '["boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg",]' )
96
1
"""simple docstring""" def _snake_case ( lowercase__ ): assert column_title.isupper() _lowerCamelCase : List[str] = 0 _lowerCamelCase : Any = len(lowercase__ ) - 1 _lowerCamelCase : Optional[Any] = 0 while index >= 0: _lowerCamelCase : Union[str, Any] = (ord(column_title[index] ) - 64) * pow(26 , lowercase__ ) answer += value power += 1 index -= 1 return answer if __name__ == "__main__": from doctest import testmod testmod()
96
"""simple docstring""" import unittest import numpy as np import torch from diffusers import DDIMPipeline, DDIMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow, torch_device from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class lowerCAmelCase__ ( lowercase, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = DDIMPipeline lowerCamelCase__ = UNCONDITIONAL_IMAGE_GENERATION_PARAMS lowerCamelCase__ = PipelineTesterMixin.required_optional_params - { """num_images_per_prompt""", """latents""", """callback""", """callback_steps""", } lowerCamelCase__ = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS lowerCamelCase__ = False def A_ ( self ): torch.manual_seed(0 ) _lowerCamelCase : List[Any] = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('DownBlock2D', 'AttnDownBlock2D') , up_block_types=('AttnUpBlock2D', 'UpBlock2D') , ) _lowerCamelCase : List[str] = DDIMScheduler() _lowerCamelCase : Optional[int] = {'unet': unet, 'scheduler': scheduler} return components def A_ ( self , lowercase , lowercase=0 ): if str(lowercase ).startswith('mps' ): _lowerCamelCase : Dict = torch.manual_seed(lowercase ) else: _lowerCamelCase : List[str] = torch.Generator(device=lowercase ).manual_seed(lowercase ) _lowerCamelCase : Tuple = { 'batch_size': 1, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs def A_ ( self ): _lowerCamelCase : Any = 'cpu' _lowerCamelCase : Tuple = self.get_dummy_components() _lowerCamelCase : Optional[Any] = self.pipeline_class(**lowercase ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : str = self.get_dummy_inputs(lowercase ) _lowerCamelCase : int = pipe(**lowercase ).images _lowerCamelCase : Any = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 32, 32, 3) ) _lowerCamelCase : Tuple = np.array( [1.000E00, 5.717E-01, 4.717E-01, 1.000E00, 0.000E00, 1.000E00, 3.000E-04, 0.000E00, 9.000E-04] ) _lowerCamelCase : str = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowercase , 1E-3 ) def A_ ( self ): super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) def A_ ( self ): super().test_save_load_local(expected_max_difference=3E-3 ) def A_ ( self ): super().test_save_load_optional_components(expected_max_difference=3E-3 ) def A_ ( self ): super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def A_ ( self ): _lowerCamelCase : Optional[Any] = 'google/ddpm-cifar10-32' _lowerCamelCase : Optional[Any] = UNetaDModel.from_pretrained(lowercase ) _lowerCamelCase : Dict = DDIMScheduler() _lowerCamelCase : Dict = DDIMPipeline(unet=lowercase , scheduler=lowercase ) ddim.to(lowercase ) ddim.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : List[str] = torch.manual_seed(0 ) _lowerCamelCase : str = ddim(generator=lowercase , eta=0.0 , output_type='numpy' ).images _lowerCamelCase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _lowerCamelCase : List[Any] = np.array([0.17_23, 0.16_17, 0.16_00, 0.16_26, 0.14_97, 0.15_13, 0.15_05, 0.14_42, 0.14_53] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def A_ ( self ): _lowerCamelCase : Optional[int] = 'google/ddpm-ema-bedroom-256' _lowerCamelCase : str = UNetaDModel.from_pretrained(lowercase ) _lowerCamelCase : str = DDIMScheduler.from_pretrained(lowercase ) _lowerCamelCase : Optional[int] = DDIMPipeline(unet=lowercase , scheduler=lowercase ) ddpm.to(lowercase ) ddpm.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : Tuple = torch.manual_seed(0 ) _lowerCamelCase : int = ddpm(generator=lowercase , output_type='numpy' ).images _lowerCamelCase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _lowerCamelCase : str = np.array([0.00_60, 0.02_01, 0.03_44, 0.00_24, 0.00_18, 0.00_02, 0.00_22, 0.00_00, 0.00_69] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
96
1
"""simple docstring""" # This script creates a super tiny model that is useful inside tests, when we just want to test that # the machinery works, without needing to the check the quality of the outcomes. # # This version creates a tiny vocab first, and then a tiny model - so the outcome is truly tiny - # all files ~60KB. As compared to taking a full-size model, reducing to the minimum its layers and # emb dimensions, but keeping the full vocab + merges files, leading to ~3MB in total for all files. # The latter is done by `fsmt-make-super-tiny-model.py`. # # It will be used then as "stas/tiny-wmt19-en-ru" from pathlib import Path import json import tempfile from transformers import FSMTTokenizer, FSMTConfig, FSMTForConditionalGeneration from transformers.models.fsmt.tokenization_fsmt import VOCAB_FILES_NAMES lowercase__ = """tiny-wmt19-en-ru""" # Build # borrowed from a test lowercase__ = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """w</w>""", """r</w>""", """t</w>""", """lo""", """low""", """er</w>""", """low</w>""", """lowest</w>""", """newer</w>""", """wider</w>""", """<unk>""", ] lowercase__ = dict(zip(vocab, range(len(vocab)))) lowercase__ = ["""l o 123""", """lo w 1456""", """e r</w> 1789""", """"""] with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ = Path(tmpdirname) lowercase__ = build_dir / VOCAB_FILES_NAMES["""src_vocab_file"""] lowercase__ = build_dir / VOCAB_FILES_NAMES["""tgt_vocab_file"""] lowercase__ = build_dir / VOCAB_FILES_NAMES["""merges_file"""] with open(src_vocab_file, """w""") as fp: fp.write(json.dumps(vocab_tokens)) with open(tgt_vocab_file, """w""") as fp: fp.write(json.dumps(vocab_tokens)) with open(merges_file, """w""") as fp: fp.write("""\n""".join(merges)) lowercase__ = FSMTTokenizer( langs=["""en""", """ru"""], src_vocab_size=len(vocab), tgt_vocab_size=len(vocab), src_vocab_file=src_vocab_file, tgt_vocab_file=tgt_vocab_file, merges_file=merges_file, ) lowercase__ = FSMTConfig( langs=["""ru""", """en"""], src_vocab_size=1000, tgt_vocab_size=1000, d_model=4, encoder_layers=1, decoder_layers=1, encoder_ffn_dim=4, decoder_ffn_dim=4, encoder_attention_heads=1, decoder_attention_heads=1, ) lowercase__ = FSMTForConditionalGeneration(config) print(F"num of params {tiny_model.num_parameters()}") # Test lowercase__ = tokenizer(["""Making tiny model"""], return_tensors="""pt""") lowercase__ = tiny_model(**batch) print("""test output:""", len(outputs.logits[0])) # Save tiny_model.half() # makes it smaller tiny_model.save_pretrained(mname_tiny) tokenizer.save_pretrained(mname_tiny) print(F"Generated {mname_tiny}") # Upload # transformers-cli upload tiny-wmt19-en-ru
96
"""simple docstring""" # Imports import numpy as np class lowerCAmelCase__ : '''simple docstring''' def __init__( self , lowercase=None , lowercase=None , lowercase=None , lowercase=None , lowercase=None ): self.set_matricies(red=lowercase , green=lowercase , blue=lowercase , red_edge=lowercase , nir=lowercase ) def A_ ( self , lowercase=None , lowercase=None , lowercase=None , lowercase=None , lowercase=None ): if red is not None: _lowerCamelCase : Optional[int] = red if green is not None: _lowerCamelCase : Optional[Any] = green if blue is not None: _lowerCamelCase : Tuple = blue if red_edge is not None: _lowerCamelCase : Optional[Any] = red_edge if nir is not None: _lowerCamelCase : Union[str, Any] = nir return True def A_ ( self , lowercase="" , lowercase=None , lowercase=None , lowercase=None , lowercase=None , lowercase=None ): self.set_matricies(red=lowercase , green=lowercase , blue=lowercase , red_edge=lowercase , nir=lowercase ) _lowerCamelCase : str = { 'ARVI2': self.arvaa, 'CCCI': self.ccci, 'CVI': self.cvi, 'GLI': self.gli, 'NDVI': self.ndvi, 'BNDVI': self.bndvi, 'redEdgeNDVI': self.red_edge_ndvi, 'GNDVI': self.gndvi, 'GBNDVI': self.gbndvi, 'GRNDVI': self.grndvi, 'RBNDVI': self.rbndvi, 'PNDVI': self.pndvi, 'ATSAVI': self.atsavi, 'BWDRVI': self.bwdrvi, 'CIgreen': self.ci_green, 'CIrededge': self.ci_rededge, 'CI': self.ci, 'CTVI': self.ctvi, 'GDVI': self.gdvi, 'EVI': self.evi, 'GEMI': self.gemi, 'GOSAVI': self.gosavi, 'GSAVI': self.gsavi, 'Hue': self.hue, 'IVI': self.ivi, 'IPVI': self.ipvi, 'I': self.i, 'RVI': self.rvi, 'MRVI': self.mrvi, 'MSAVI': self.m_savi, 'NormG': self.norm_g, 'NormNIR': self.norm_nir, 'NormR': self.norm_r, 'NGRDI': self.ngrdi, 'RI': self.ri, 'S': self.s, 'IF': self._if, 'DVI': self.dvi, 'TVI': self.tvi, 'NDRE': self.ndre, } try: return funcs[index]() except KeyError: print('Index not in the list!' ) return False def A_ ( self ): return -0.18 + (1.17 * ((self.nir - self.red) / (self.nir + self.red))) def A_ ( self ): return ((self.nir - self.redEdge) / (self.nir + self.redEdge)) / ( (self.nir - self.red) / (self.nir + self.red) ) def A_ ( self ): return self.nir * (self.red / (self.green**2)) def A_ ( self ): return (2 * self.green - self.red - self.blue) / ( 2 * self.green + self.red + self.blue ) def A_ ( self ): return (self.nir - self.red) / (self.nir + self.red) def A_ ( self ): return (self.nir - self.blue) / (self.nir + self.blue) def A_ ( self ): return (self.redEdge - self.red) / (self.redEdge + self.red) def A_ ( self ): return (self.nir - self.green) / (self.nir + self.green) def A_ ( self ): return (self.nir - (self.green + self.blue)) / ( self.nir + (self.green + self.blue) ) def A_ ( self ): return (self.nir - (self.green + self.red)) / ( self.nir + (self.green + self.red) ) def A_ ( self ): return (self.nir - (self.blue + self.red)) / (self.nir + (self.blue + self.red)) def A_ ( self ): return (self.nir - (self.green + self.red + self.blue)) / ( self.nir + (self.green + self.red + self.blue) ) def A_ ( self , lowercase=0.08 , lowercase=1.22 , lowercase=0.03 ): return a * ( (self.nir - a * self.red - b) / (a * self.nir + self.red - a * b + x * (1 + a**2)) ) def A_ ( self ): return (0.1 * self.nir - self.blue) / (0.1 * self.nir + self.blue) def A_ ( self ): return (self.nir / self.green) - 1 def A_ ( self ): return (self.nir / self.redEdge) - 1 def A_ ( self ): return (self.red - self.blue) / self.red def A_ ( self ): _lowerCamelCase : Any = self.ndvi() return ((ndvi + 0.5) / (abs(ndvi + 0.5 ))) * (abs(ndvi + 0.5 ) ** (1 / 2)) def A_ ( self ): return self.nir - self.green def A_ ( self ): return 2.5 * ( (self.nir - self.red) / (self.nir + 6 * self.red - 7.5 * self.blue + 1) ) def A_ ( self ): _lowerCamelCase : Any = (2 * (self.nir**2 - self.red**2) + 1.5 * self.nir + 0.5 * self.red) / ( self.nir + self.red + 0.5 ) return n * (1 - 0.25 * n) - (self.red - 0.1_25) / (1 - self.red) def A_ ( self , lowercase=0.16 ): return (self.nir - self.green) / (self.nir + self.green + y) def A_ ( self , lowercase=0.5 ): return ((self.nir - self.green) / (self.nir + self.green + n)) * (1 + n) def A_ ( self ): return np.arctan( ((2 * self.red - self.green - self.blue) / 30.5) * (self.green - self.blue) ) def A_ ( self , lowercase=None , lowercase=None ): return (self.nir - b) / (a * self.red) def A_ ( self ): return (self.nir / ((self.nir + self.red) / 2)) * (self.ndvi() + 1) def A_ ( self ): return (self.red + self.green + self.blue) / 30.5 def A_ ( self ): return self.nir / self.red def A_ ( self ): return (self.rvi() - 1) / (self.rvi() + 1) def A_ ( self ): return ( (2 * self.nir + 1) - ((2 * self.nir + 1) ** 2 - 8 * (self.nir - self.red)) ** (1 / 2) ) / 2 def A_ ( self ): return self.green / (self.nir + self.red + self.green) def A_ ( self ): return self.nir / (self.nir + self.red + self.green) def A_ ( self ): return self.red / (self.nir + self.red + self.green) def A_ ( self ): return (self.green - self.red) / (self.green + self.red) def A_ ( self ): return (self.red - self.green) / (self.red + self.green) def A_ ( self ): _lowerCamelCase : Union[str, Any] = np.max([np.max(self.red ), np.max(self.green ), np.max(self.blue )] ) _lowerCamelCase : Dict = np.min([np.min(self.red ), np.min(self.green ), np.min(self.blue )] ) return (max_value - min_value) / max_value def A_ ( self ): return (2 * self.red - self.green - self.blue) / (self.green - self.blue) def A_ ( self ): return self.nir / self.red def A_ ( self ): return (self.ndvi() + 0.5) ** (1 / 2) def A_ ( self ): return (self.nir - self.redEdge) / (self.nir + self.redEdge)
96
1
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = {"""vocab_file""": """spm_char.model"""} lowercase__ = { """vocab_file""": { """microsoft/speecht5_asr""": """https://huggingface.co/microsoft/speecht5_asr/resolve/main/spm_char.model""", """microsoft/speecht5_tts""": """https://huggingface.co/microsoft/speecht5_tts/resolve/main/spm_char.model""", """microsoft/speecht5_vc""": """https://huggingface.co/microsoft/speecht5_vc/resolve/main/spm_char.model""", } } lowercase__ = { """microsoft/speecht5_asr""": 1024, """microsoft/speecht5_tts""": 1024, """microsoft/speecht5_vc""": 1024, } class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = VOCAB_FILES_NAMES lowerCamelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ = ["""input_ids""", """attention_mask"""] def __init__( self , lowercase , lowercase="<s>" , lowercase="</s>" , lowercase="<unk>" , lowercase="<pad>" , lowercase = None , **lowercase , ): _lowerCamelCase : List[str] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=lowercase , eos_token=lowercase , unk_token=lowercase , pad_token=lowercase , sp_model_kwargs=self.sp_model_kwargs , **lowercase , ) _lowerCamelCase : str = vocab_file _lowerCamelCase : Union[str, Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowercase ) @property def A_ ( self ): return self.sp_model.get_piece_size() def A_ ( self ): _lowerCamelCase : Optional[int] = {self.convert_ids_to_tokens(lowercase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): _lowerCamelCase : Optional[Any] = self.__dict__.copy() _lowerCamelCase : Tuple = None return state def __setstate__( self , lowercase ): _lowerCamelCase : int = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): _lowerCamelCase : Tuple = {} _lowerCamelCase : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def A_ ( self , lowercase ): return self.sp_model.encode(lowercase , out_type=lowercase ) def A_ ( self , lowercase ): return self.sp_model.piece_to_id(lowercase ) def A_ ( self , lowercase ): _lowerCamelCase : Dict = self.sp_model.IdToPiece(lowercase ) return token def A_ ( self , lowercase ): _lowerCamelCase : Dict = [] _lowerCamelCase : List[str] = '' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(lowercase ) + token _lowerCamelCase : List[Any] = [] else: current_sub_tokens.append(lowercase ) out_string += self.sp_model.decode(lowercase ) return out_string.strip() def A_ ( self , lowercase , lowercase=None ): if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def A_ ( self , lowercase , lowercase = None , lowercase = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowercase , token_ids_a=lowercase , already_has_special_tokens=lowercase ) _lowerCamelCase : Optional[int] = [1] if token_ids_a is None: return ([0] * len(lowercase )) + suffix_ones return ([0] * len(lowercase )) + ([0] * len(lowercase )) + suffix_ones def A_ ( self , lowercase , lowercase = None ): if not os.path.isdir(lowercase ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return _lowerCamelCase : Tuple = os.path.join( lowercase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowercase ) elif not os.path.isfile(self.vocab_file ): with open(lowercase , 'wb' ) as fi: _lowerCamelCase : str = self.sp_model.serialized_model_proto() fi.write(lowercase ) return (out_vocab_file,)
96
"""simple docstring""" import torch from torch import nn from transformers import CLIPPreTrainedModel, CLIPVisionModel from ...models.attention import BasicTransformerBlock from ...utils import logging lowercase__ = logging.get_logger(__name__) # pylint: disable=invalid-name class lowerCAmelCase__ ( lowercase ): '''simple docstring''' def __init__( self , lowercase , lowercase=768 ): super().__init__(lowercase ) _lowerCamelCase : Any = proj_size _lowerCamelCase : Dict = CLIPVisionModel(lowercase ) _lowerCamelCase : List[str] = PaintByExampleMapper(lowercase ) _lowerCamelCase : Optional[Any] = nn.LayerNorm(config.hidden_size ) _lowerCamelCase : int = nn.Linear(config.hidden_size , self.proj_size ) # uncondition for scaling _lowerCamelCase : str = nn.Parameter(torch.randn((1, 1, self.proj_size) ) ) def A_ ( self , lowercase , lowercase=False ): _lowerCamelCase : Union[str, Any] = self.model(pixel_values=lowercase ) _lowerCamelCase : int = clip_output.pooler_output _lowerCamelCase : str = self.mapper(latent_states[:, None] ) _lowerCamelCase : List[Any] = self.final_layer_norm(lowercase ) _lowerCamelCase : Dict = self.proj_out(lowercase ) if return_uncond_vector: return latent_states, self.uncond_vector return latent_states class lowerCAmelCase__ ( nn.Module ): '''simple docstring''' def __init__( self , lowercase ): super().__init__() _lowerCamelCase : Tuple = (config.num_hidden_layers + 1) // 5 _lowerCamelCase : int = config.hidden_size _lowerCamelCase : Optional[Any] = 1 _lowerCamelCase : str = nn.ModuleList( [ BasicTransformerBlock(lowercase , lowercase , lowercase , activation_fn='gelu' , attention_bias=lowercase ) for _ in range(lowercase ) ] ) def A_ ( self , lowercase ): for block in self.blocks: _lowerCamelCase : Tuple = block(lowercase ) return hidden_states
96
1
"""simple docstring""" import gc import unittest import numpy as np import torch from diffusers import StableDiffusionKDiffusionPipeline from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() @slow @require_torch_gpu class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def A_ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A_ ( self ): _lowerCamelCase : Tuple = StableDiffusionKDiffusionPipeline.from_pretrained('CompVis/stable-diffusion-v1-4' ) _lowerCamelCase : Any = sd_pipe.to(lowercase ) sd_pipe.set_progress_bar_config(disable=lowercase ) sd_pipe.set_scheduler('sample_euler' ) _lowerCamelCase : Dict = 'A painting of a squirrel eating a burger' _lowerCamelCase : List[Any] = torch.manual_seed(0 ) _lowerCamelCase : Dict = sd_pipe([prompt] , generator=lowercase , guidance_scale=9.0 , num_inference_steps=20 , output_type='np' ) _lowerCamelCase : Union[str, Any] = output.images _lowerCamelCase : Any = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _lowerCamelCase : Dict = np.array([0.04_47, 0.04_92, 0.04_68, 0.04_08, 0.03_83, 0.04_08, 0.03_54, 0.03_80, 0.03_39] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def A_ ( self ): _lowerCamelCase : Any = StableDiffusionKDiffusionPipeline.from_pretrained('stabilityai/stable-diffusion-2-1-base' ) _lowerCamelCase : Any = sd_pipe.to(lowercase ) sd_pipe.set_progress_bar_config(disable=lowercase ) sd_pipe.set_scheduler('sample_euler' ) _lowerCamelCase : Union[str, Any] = 'A painting of a squirrel eating a burger' _lowerCamelCase : List[str] = torch.manual_seed(0 ) _lowerCamelCase : Tuple = sd_pipe([prompt] , generator=lowercase , guidance_scale=9.0 , num_inference_steps=20 , output_type='np' ) _lowerCamelCase : Optional[int] = output.images _lowerCamelCase : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _lowerCamelCase : List[str] = np.array([0.12_37, 0.13_20, 0.14_38, 0.13_59, 0.13_90, 0.11_32, 0.12_77, 0.11_75, 0.11_12] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-1 def A_ ( self ): _lowerCamelCase : Union[str, Any] = StableDiffusionKDiffusionPipeline.from_pretrained('stabilityai/stable-diffusion-2-1-base' ) _lowerCamelCase : int = sd_pipe.to(lowercase ) sd_pipe.set_progress_bar_config(disable=lowercase ) sd_pipe.set_scheduler('sample_dpmpp_2m' ) _lowerCamelCase : int = 'A painting of a squirrel eating a burger' _lowerCamelCase : List[str] = torch.manual_seed(0 ) _lowerCamelCase : Any = sd_pipe( [prompt] , generator=lowercase , guidance_scale=7.5 , num_inference_steps=15 , output_type='np' , use_karras_sigmas=lowercase , ) _lowerCamelCase : List[Any] = output.images _lowerCamelCase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _lowerCamelCase : Any = np.array( [0.11_38_16_89, 0.12_11_29_21, 0.1_38_94_57, 0.12_54_96_06, 0.1_24_49_64, 0.10_83_15_17, 0.11_56_28_66, 0.10_86_78_16, 0.10_49_90_48] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
96
"""simple docstring""" lowercase__ = { """meter""": """m""", """kilometer""": """km""", """megametre""": """Mm""", """gigametre""": """Gm""", """terametre""": """Tm""", """petametre""": """Pm""", """exametre""": """Em""", """zettametre""": """Zm""", """yottametre""": """Ym""", } # Exponent of the factor(meter) lowercase__ = { """m""": 0, """km""": 3, """Mm""": 6, """Gm""": 9, """Tm""": 12, """Pm""": 15, """Em""": 18, """Zm""": 21, """Ym""": 24, } def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : List[Any] = from_type.lower().strip('s' ) _lowerCamelCase : List[Any] = to_type.lower().strip('s' ) _lowerCamelCase : Optional[int] = UNIT_SYMBOL.get(lowercase__ , lowercase__ ) _lowerCamelCase : Any = UNIT_SYMBOL.get(lowercase__ , lowercase__ ) if from_sanitized not in METRIC_CONVERSION: _lowerCamelCase : Tuple = ( f'''Invalid \'from_type\' value: {from_type!r}.\n''' f'''Conversion abbreviations are: {', '.join(lowercase__ )}''' ) raise ValueError(lowercase__ ) if to_sanitized not in METRIC_CONVERSION: _lowerCamelCase : Any = ( f'''Invalid \'to_type\' value: {to_type!r}.\n''' f'''Conversion abbreviations are: {', '.join(lowercase__ )}''' ) raise ValueError(lowercase__ ) _lowerCamelCase : List[Any] = METRIC_CONVERSION[from_sanitized] _lowerCamelCase : int = METRIC_CONVERSION[to_sanitized] _lowerCamelCase : List[str] = 1 if from_exponent > to_exponent: _lowerCamelCase : List[str] = from_exponent - to_exponent else: _lowerCamelCase : List[Any] = -(to_exponent - from_exponent) return value * pow(10 , lowercase__ ) if __name__ == "__main__": from doctest import testmod testmod()
96
1
"""simple docstring""" import torch from diffusers import DDPMScheduler from .test_schedulers import SchedulerCommonTest class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = (DDPMScheduler,) def A_ ( self , **lowercase ): _lowerCamelCase : Dict = { 'num_train_timesteps': 1000, 'beta_start': 0.00_01, 'beta_end': 0.02, 'beta_schedule': 'linear', 'variance_type': 'fixed_small', 'clip_sample': True, } config.update(**lowercase ) return config def A_ ( self ): for timesteps in [1, 5, 100, 1000]: self.check_over_configs(num_train_timesteps=lowercase ) def A_ ( self ): for beta_start, beta_end in zip([0.00_01, 0.0_01, 0.01, 0.1] , [0.0_02, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=lowercase , beta_end=lowercase ) def A_ ( self ): for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=lowercase ) def A_ ( self ): for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=lowercase ) def A_ ( self ): for clip_sample in [True, False]: self.check_over_configs(clip_sample=lowercase ) def A_ ( self ): self.check_over_configs(thresholding=lowercase ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=lowercase , prediction_type=lowercase , sample_max_value=lowercase , ) def A_ ( self ): for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=lowercase ) def A_ ( self ): for t in [0, 500, 999]: self.check_over_forward(time_step=lowercase ) def A_ ( self ): _lowerCamelCase : Union[str, Any] = self.scheduler_classes[0] _lowerCamelCase : int = self.get_scheduler_config() _lowerCamelCase : List[Any] = scheduler_class(**lowercase ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0_09_79 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.02 ) ) < 1E-5 def A_ ( self ): _lowerCamelCase : Optional[int] = self.scheduler_classes[0] _lowerCamelCase : int = self.get_scheduler_config() _lowerCamelCase : Optional[int] = scheduler_class(**lowercase ) _lowerCamelCase : Tuple = len(lowercase ) _lowerCamelCase : int = self.dummy_model() _lowerCamelCase : Tuple = self.dummy_sample_deter _lowerCamelCase : Tuple = torch.manual_seed(0 ) for t in reversed(range(lowercase ) ): # 1. predict noise residual _lowerCamelCase : Tuple = model(lowercase , lowercase ) # 2. predict previous mean of sample x_t-1 _lowerCamelCase : str = scheduler.step(lowercase , lowercase , lowercase , generator=lowercase ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance _lowerCamelCase : List[str] = pred_prev_sample _lowerCamelCase : List[Any] = torch.sum(torch.abs(lowercase ) ) _lowerCamelCase : Dict = torch.mean(torch.abs(lowercase ) ) assert abs(result_sum.item() - 2_58.96_06 ) < 1E-2 assert abs(result_mean.item() - 0.33_72 ) < 1E-3 def A_ ( self ): _lowerCamelCase : str = self.scheduler_classes[0] _lowerCamelCase : int = self.get_scheduler_config(prediction_type='v_prediction' ) _lowerCamelCase : List[Any] = scheduler_class(**lowercase ) _lowerCamelCase : List[Any] = len(lowercase ) _lowerCamelCase : int = self.dummy_model() _lowerCamelCase : Any = self.dummy_sample_deter _lowerCamelCase : Any = torch.manual_seed(0 ) for t in reversed(range(lowercase ) ): # 1. predict noise residual _lowerCamelCase : Union[str, Any] = model(lowercase , lowercase ) # 2. predict previous mean of sample x_t-1 _lowerCamelCase : Any = scheduler.step(lowercase , lowercase , lowercase , generator=lowercase ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance _lowerCamelCase : Optional[int] = pred_prev_sample _lowerCamelCase : List[str] = torch.sum(torch.abs(lowercase ) ) _lowerCamelCase : Union[str, Any] = torch.mean(torch.abs(lowercase ) ) assert abs(result_sum.item() - 2_02.02_96 ) < 1E-2 assert abs(result_mean.item() - 0.26_31 ) < 1E-3 def A_ ( self ): _lowerCamelCase : List[Any] = self.scheduler_classes[0] _lowerCamelCase : List[str] = self.get_scheduler_config() _lowerCamelCase : Any = scheduler_class(**lowercase ) _lowerCamelCase : Dict = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=lowercase ) _lowerCamelCase : Any = scheduler.timesteps for i, timestep in enumerate(lowercase ): if i == len(lowercase ) - 1: _lowerCamelCase : Optional[int] = -1 else: _lowerCamelCase : Optional[Any] = timesteps[i + 1] _lowerCamelCase : Optional[int] = scheduler.previous_timestep(lowercase ) _lowerCamelCase : Tuple = prev_t.item() self.assertEqual(lowercase , lowercase ) def A_ ( self ): _lowerCamelCase : Union[str, Any] = self.scheduler_classes[0] _lowerCamelCase : str = self.get_scheduler_config() _lowerCamelCase : Tuple = scheduler_class(**lowercase ) _lowerCamelCase : List[Any] = [100, 87, 50, 51, 0] with self.assertRaises(lowercase , msg='`custom_timesteps` must be in descending order.' ): scheduler.set_timesteps(timesteps=lowercase ) def A_ ( self ): _lowerCamelCase : str = self.scheduler_classes[0] _lowerCamelCase : List[Any] = self.get_scheduler_config() _lowerCamelCase : int = scheduler_class(**lowercase ) _lowerCamelCase : int = [100, 87, 50, 1, 0] _lowerCamelCase : Optional[int] = len(lowercase ) with self.assertRaises(lowercase , msg='Can only pass one of `num_inference_steps` or `custom_timesteps`.' ): scheduler.set_timesteps(num_inference_steps=lowercase , timesteps=lowercase ) def A_ ( self ): _lowerCamelCase : Optional[int] = self.scheduler_classes[0] _lowerCamelCase : int = self.get_scheduler_config() _lowerCamelCase : List[Any] = scheduler_class(**lowercase ) _lowerCamelCase : int = [scheduler.config.num_train_timesteps] with self.assertRaises( lowercase , msg='`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}' , ): scheduler.set_timesteps(timesteps=lowercase )
96
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ = {"""configuration_vit_msn""": ["""VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ViTMSNConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ """VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST""", """ViTMSNModel""", """ViTMSNForImageClassification""", """ViTMSNPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_vit_msn import VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMSNConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_msn import ( VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMSNForImageClassification, ViTMSNModel, ViTMSNPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
96
1
"""simple docstring""" import math def _snake_case ( lowercase__ ): return math.sqrt(lowercase__ ) * math.sqrt(lowercase__ ) == num def _snake_case ( lowercase__ ): _lowerCamelCase : Optional[int] = 0 _lowerCamelCase : List[Any] = n while left <= right: _lowerCamelCase : str = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: _lowerCamelCase : str = mid - 1 else: _lowerCamelCase : Optional[int] = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
96
"""simple docstring""" def _snake_case ( lowercase__ , lowercase__ ): # "extended trapezoidal rule" # int(f) = dx/2 * (f1 + 2f2 + ... + fn) _lowerCamelCase : List[Any] = (boundary[1] - boundary[0]) / steps _lowerCamelCase : Tuple = boundary[0] _lowerCamelCase : Dict = boundary[1] _lowerCamelCase : List[Any] = make_points(lowercase__ , lowercase__ , lowercase__ ) _lowerCamelCase : List[Any] = 0.0 y += (h / 2.0) * f(lowercase__ ) for i in x_i: # print(i) y += h * f(lowercase__ ) y += (h / 2.0) * f(lowercase__ ) return y def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : str = a + h while x < (b - h): yield x _lowerCamelCase : int = x + h def _snake_case ( lowercase__ ): # enter your function here _lowerCamelCase : Optional[Any] = (x - 0) * (x - 0) return y def _snake_case ( ): _lowerCamelCase : int = 0.0 # Lower bound of integration _lowerCamelCase : Optional[int] = 1.0 # Upper bound of integration _lowerCamelCase : List[str] = 1_0.0 # define number of steps or resolution _lowerCamelCase : List[Any] = [a, b] # define boundary of integration _lowerCamelCase : Optional[Any] = method_a(lowercase__ , lowercase__ ) print(f'''y = {y}''' ) if __name__ == "__main__": main()
96
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowercase__ = { """configuration_poolformer""": [ """POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """PoolFormerConfig""", """PoolFormerOnnxConfig""", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ["""PoolFormerFeatureExtractor"""] lowercase__ = ["""PoolFormerImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ """POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """PoolFormerForImageClassification""", """PoolFormerModel""", """PoolFormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
96
"""simple docstring""" import math def _snake_case ( lowercase__ ): return math.sqrt(lowercase__ ) * math.sqrt(lowercase__ ) == num def _snake_case ( lowercase__ ): _lowerCamelCase : Optional[int] = 0 _lowerCamelCase : List[Any] = n while left <= right: _lowerCamelCase : str = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: _lowerCamelCase : str = mid - 1 else: _lowerCamelCase : Optional[int] = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" from typing import List from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { """snap-research/efficientformer-l1-300""": ( """https://huggingface.co/snap-research/efficientformer-l1-300/resolve/main/config.json""" ), } class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """efficientformer""" def __init__( self , lowercase = [3, 2, 6, 4] , lowercase = [48, 96, 224, 448] , lowercase = [True, True, True, True] , lowercase = 448 , lowercase = 32 , lowercase = 4 , lowercase = 7 , lowercase = 5 , lowercase = 8 , lowercase = 4 , lowercase = 0.0 , lowercase = 16 , lowercase = 3 , lowercase = 3 , lowercase = 3 , lowercase = 2 , lowercase = 1 , lowercase = 0.0 , lowercase = 1 , lowercase = True , lowercase = True , lowercase = 1E-5 , lowercase = "gelu" , lowercase = 0.02 , lowercase = 1E-12 , lowercase = 224 , lowercase = 1E-05 , **lowercase , ): super().__init__(**lowercase ) _lowerCamelCase : Dict = hidden_act _lowerCamelCase : Optional[Any] = hidden_dropout_prob _lowerCamelCase : Optional[Any] = hidden_sizes _lowerCamelCase : Optional[Any] = num_hidden_layers _lowerCamelCase : int = num_attention_heads _lowerCamelCase : Tuple = initializer_range _lowerCamelCase : Tuple = layer_norm_eps _lowerCamelCase : Optional[Any] = patch_size _lowerCamelCase : List[str] = num_channels _lowerCamelCase : List[Any] = depths _lowerCamelCase : List[str] = mlp_expansion_ratio _lowerCamelCase : Union[str, Any] = downsamples _lowerCamelCase : List[str] = dim _lowerCamelCase : str = key_dim _lowerCamelCase : Tuple = attention_ratio _lowerCamelCase : Dict = resolution _lowerCamelCase : Any = pool_size _lowerCamelCase : Any = downsample_patch_size _lowerCamelCase : str = downsample_stride _lowerCamelCase : Tuple = downsample_pad _lowerCamelCase : List[str] = drop_path_rate _lowerCamelCase : Optional[Any] = num_metaad_blocks _lowerCamelCase : Optional[Any] = distillation _lowerCamelCase : int = use_layer_scale _lowerCamelCase : Optional[int] = layer_scale_init_value _lowerCamelCase : List[Any] = image_size _lowerCamelCase : Optional[Any] = batch_norm_eps
96
"""simple docstring""" import functools from typing import Any def _snake_case ( lowercase__ , lowercase__ ): # Validation if not isinstance(lowercase__ , lowercase__ ) or len(lowercase__ ) == 0: raise ValueError('the string should be not empty string' ) if not isinstance(lowercase__ , lowercase__ ) or not all( isinstance(lowercase__ , lowercase__ ) and len(lowercase__ ) > 0 for item in words ): raise ValueError('the words should be a list of non-empty strings' ) # Build trie _lowerCamelCase : dict[str, Any] = {} _lowerCamelCase : List[Any] = 'WORD_KEEPER' for word in words: _lowerCamelCase : Dict = trie for c in word: if c not in trie_node: _lowerCamelCase : Any = {} _lowerCamelCase : str = trie_node[c] _lowerCamelCase : Optional[Any] = True _lowerCamelCase : Dict = len(lowercase__ ) # Dynamic programming method @functools.cache def is_breakable(lowercase__ ) -> bool: if index == len_string: return True _lowerCamelCase : List[Any] = trie for i in range(lowercase__ , lowercase__ ): _lowerCamelCase : Any = trie_node.get(string[i] , lowercase__ ) if trie_node is None: return False if trie_node.get(lowercase__ , lowercase__ ) and is_breakable(i + 1 ): return True return False return is_breakable(0 ) if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { """EleutherAI/gpt-neo-1.3B""": """https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json""", # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """gpt_neo""" lowerCamelCase__ = ["""past_key_values"""] lowerCamelCase__ = {"""num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers"""} def __init__( self , lowercase=50257 , lowercase=2048 , lowercase=2048 , lowercase=24 , lowercase=[[["global", "local"], 12]] , lowercase=16 , lowercase=None , lowercase=256 , lowercase="gelu_new" , lowercase=0.0 , lowercase=0.0 , lowercase=0.0 , lowercase=0.1 , lowercase=1E-5 , lowercase=0.02 , lowercase=True , lowercase=50256 , lowercase=50256 , **lowercase , ): _lowerCamelCase : str = vocab_size _lowerCamelCase : Union[str, Any] = max_position_embeddings _lowerCamelCase : Optional[Any] = hidden_size _lowerCamelCase : List[str] = num_layers _lowerCamelCase : Optional[Any] = num_heads _lowerCamelCase : Tuple = intermediate_size _lowerCamelCase : Optional[int] = window_size _lowerCamelCase : int = activation_function _lowerCamelCase : Optional[Any] = resid_dropout _lowerCamelCase : Any = embed_dropout _lowerCamelCase : Any = attention_dropout _lowerCamelCase : Optional[int] = classifier_dropout _lowerCamelCase : List[str] = layer_norm_epsilon _lowerCamelCase : Dict = initializer_range _lowerCamelCase : Optional[int] = use_cache _lowerCamelCase : List[str] = bos_token_id _lowerCamelCase : int = eos_token_id _lowerCamelCase : int = attention_types _lowerCamelCase : List[str] = self.expand_attention_types_params(lowercase ) if len(self.attention_layers ) != self.num_layers: raise ValueError( 'Configuration for convolutional module is incorrect. ' 'It is required that `len(config.attention_layers)` == `config.num_layers` ' F'''but is `len(config.attention_layers) = {len(self.attention_layers )}`, ''' F'''`config.num_layers = {self.num_layers}`. ''' '`config.attention_layers` is prepared using `config.attention_types`. ' 'Please verify the value of `config.attention_types` argument.' ) super().__init__(bos_token_id=lowercase , eos_token_id=lowercase , **lowercase ) @staticmethod def A_ ( lowercase ): _lowerCamelCase : Optional[int] = [] for item in attention_types: for _ in range(item[1] ): attentions.extend(item[0] ) return attentions def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): import torch _lowerCamelCase : Tuple = input.size() _lowerCamelCase : Tuple = len(lowercase__ ) _lowerCamelCase : Tuple = shape[dimension] _lowerCamelCase : Any = torch.arange(0 , lowercase__ , lowercase__ ) _lowerCamelCase : Optional[int] = torch.div(sizedim - size , lowercase__ , rounding_mode='floor' ) + 1 _lowerCamelCase : List[str] = torch.arange(lowercase__ ) + low_indices[:min_length][:, None] _lowerCamelCase : Dict = [slice(lowercase__ )] * rank _lowerCamelCase : Dict = indices _lowerCamelCase : List[Any] = input[s] _lowerCamelCase : List[Any] = list(range(0 , rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(lowercase__ ) def _snake_case ( lowercase__ , lowercase__ ): import torch _lowerCamelCase : List[str] = torch.arange(1 , lowercase__ ) _lowerCamelCase : str = torch.remainder(lowercase__ , lowercase__ ) _lowerCamelCase : Tuple = remainders == 0 _lowerCamelCase : Dict = candidates[divisor_indices] _lowerCamelCase : Union[str, Any] = torch.max(lowercase__ ) return largest_divisor, torch.div(lowercase__ , lowercase__ , rounding_mode='floor' ) class lowerCAmelCase__ ( lowercase ): '''simple docstring''' @property def A_ ( self ): _lowerCamelCase : Dict = OrderedDict({'input_ids': {0: 'batch', 1: 'sequence'}} ) if self.use_past: self.fill_with_past_key_values_(lowercase , direction='inputs' ) _lowerCamelCase : Optional[int] = {0: 'batch', 1: 'past_sequence + sequence'} else: _lowerCamelCase : Dict = {0: 'batch', 1: 'sequence'} return common_inputs @property def A_ ( self ): return self._config.num_heads def A_ ( self , lowercase , lowercase = -1 , lowercase = -1 , lowercase = False , lowercase = None , ): _lowerCamelCase : Any = super(lowercase , self ).generate_dummy_inputs( lowercase , batch_size=lowercase , seq_length=lowercase , is_pair=lowercase , framework=lowercase ) # We need to order the input in the way they appears in the forward() _lowerCamelCase : Dict = OrderedDict({'input_ids': common_inputs['input_ids']} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch _lowerCamelCase, _lowerCamelCase : List[str] = common_inputs['input_ids'].shape # Not using the same length for past_key_values _lowerCamelCase : Union[str, Any] = seqlen + 2 _lowerCamelCase : int = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) _lowerCamelCase : int = [ (torch.zeros(lowercase ), torch.zeros(lowercase )) for _ in range(self.num_layers ) ] _lowerCamelCase : List[Any] = common_inputs['attention_mask'] if self.use_past: _lowerCamelCase : int = ordered_inputs['attention_mask'].dtype _lowerCamelCase : List[str] = torch.cat( [ordered_inputs['attention_mask'], torch.ones(lowercase , lowercase , dtype=lowercase )] , dim=1 ) return ordered_inputs @property def A_ ( self ): return 13
96
"""simple docstring""" def _snake_case ( lowercase__ ): if not isinstance(lowercase__ , lowercase__ ): raise ValueError('Input series is not valid, valid series - [2, 4, 6]' ) if len(lowercase__ ) == 0: raise ValueError('Input list must be a non empty list' ) if len(lowercase__ ) == 1: return True _lowerCamelCase : List[Any] = series[1] - series[0] for index in range(len(lowercase__ ) - 1 ): if series[index + 1] - series[index] != common_diff: return False return True def _snake_case ( lowercase__ ): if not isinstance(lowercase__ , lowercase__ ): raise ValueError('Input series is not valid, valid series - [2, 4, 6]' ) if len(lowercase__ ) == 0: raise ValueError('Input list must be a non empty list' ) _lowerCamelCase : Optional[int] = 0 for val in series: answer += val return answer / len(lowercase__ ) if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" import math import time from transformers import Trainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class lowerCAmelCase__ ( lowercase ): '''simple docstring''' def __init__( self , *lowercase , lowercase=None , lowercase=None , **lowercase ): super().__init__(*lowercase , **lowercase ) _lowerCamelCase : List[str] = eval_examples _lowerCamelCase : str = post_process_function def A_ ( self , lowercase=None , lowercase=None , lowercase=None , lowercase = "eval" ): _lowerCamelCase : List[Any] = self.eval_dataset if eval_dataset is None else eval_dataset _lowerCamelCase : List[Any] = self.get_eval_dataloader(lowercase ) _lowerCamelCase : int = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. _lowerCamelCase : Optional[Any] = self.compute_metrics _lowerCamelCase : Tuple = None _lowerCamelCase : int = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop _lowerCamelCase : Any = time.time() try: _lowerCamelCase : List[str] = eval_loop( lowercase , description='Evaluation' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=lowercase , metric_key_prefix=lowercase , ) finally: _lowerCamelCase : Tuple = compute_metrics _lowerCamelCase : Dict = self.args.eval_batch_size * self.args.world_size if F'''{metric_key_prefix}_jit_compilation_time''' in output.metrics: start_time += output.metrics[F'''{metric_key_prefix}_jit_compilation_time'''] output.metrics.update( speed_metrics( lowercase , lowercase , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default _lowerCamelCase : Union[str, Any] = self.post_process_function(lowercase , lowercase , output.predictions ) _lowerCamelCase : Dict = self.compute_metrics(lowercase ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F'''{metric_key_prefix}_''' ): _lowerCamelCase : Dict = metrics.pop(lowercase ) metrics.update(output.metrics ) else: _lowerCamelCase : Tuple = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(lowercase ) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) _lowerCamelCase : Any = self.callback_handler.on_evaluate(self.args , self.state , self.control , lowercase ) return metrics def A_ ( self , lowercase , lowercase , lowercase=None , lowercase = "test" ): _lowerCamelCase : Tuple = self.get_test_dataloader(lowercase ) # Temporarily disable metric computation, we will do it in the loop here. _lowerCamelCase : Any = self.compute_metrics _lowerCamelCase : str = None _lowerCamelCase : Dict = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop _lowerCamelCase : Optional[Any] = time.time() try: _lowerCamelCase : Union[str, Any] = eval_loop( lowercase , description='Prediction' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=lowercase , metric_key_prefix=lowercase , ) finally: _lowerCamelCase : List[str] = compute_metrics _lowerCamelCase : Optional[int] = self.args.eval_batch_size * self.args.world_size if F'''{metric_key_prefix}_jit_compilation_time''' in output.metrics: start_time += output.metrics[F'''{metric_key_prefix}_jit_compilation_time'''] output.metrics.update( speed_metrics( lowercase , lowercase , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is None or self.compute_metrics is None: return output _lowerCamelCase : Any = self.post_process_function(lowercase , lowercase , output.predictions , 'predict' ) _lowerCamelCase : Dict = self.compute_metrics(lowercase ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F'''{metric_key_prefix}_''' ): _lowerCamelCase : int = metrics.pop(lowercase ) metrics.update(output.metrics ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=lowercase )
96
"""simple docstring""" import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler lowercase__ = 16 lowercase__ = 32 def _snake_case ( lowercase__ , lowercase__ = 16 , lowercase__ = "bert-base-cased" ): _lowerCamelCase : List[Any] = AutoTokenizer.from_pretrained(lowercase__ ) _lowerCamelCase : Tuple = load_dataset('glue' , 'mrpc' ) def tokenize_function(lowercase__ ): # max_length=None => use the model max length (it's actually the default) _lowerCamelCase : Union[str, Any] = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=lowercase__ , max_length=lowercase__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset _lowerCamelCase : int = datasets.map( lowercase__ , batched=lowercase__ , remove_columns=['idx', 'sentence1', 'sentence2'] , load_from_cache_file=lowercase__ ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _lowerCamelCase : Optional[int] = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(lowercase__ ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(lowercase__ , padding='max_length' , max_length=128 , return_tensors='pt' ) return tokenizer.pad(lowercase__ , padding='longest' , return_tensors='pt' ) # Instantiate dataloaders. _lowerCamelCase : List[str] = DataLoader( tokenized_datasets['train'] , shuffle=lowercase__ , collate_fn=lowercase__ , batch_size=lowercase__ ) _lowerCamelCase : int = DataLoader( tokenized_datasets['validation'] , shuffle=lowercase__ , collate_fn=lowercase__ , batch_size=lowercase__ ) return train_dataloader, eval_dataloader def _snake_case ( lowercase__ , lowercase__ ): # Initialize accelerator _lowerCamelCase : Optional[int] = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _lowerCamelCase : Optional[int] = config['lr'] _lowerCamelCase : Optional[int] = int(config['num_epochs'] ) _lowerCamelCase : Union[str, Any] = int(config['seed'] ) _lowerCamelCase : Optional[int] = int(config['batch_size'] ) _lowerCamelCase : Dict = args.model_name_or_path set_seed(lowercase__ ) _lowerCamelCase, _lowerCamelCase : Optional[int] = get_dataloaders(lowercase__ , lowercase__ , lowercase__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _lowerCamelCase : int = AutoModelForSequenceClassification.from_pretrained(lowercase__ , return_dict=lowercase__ ) # Instantiate optimizer _lowerCamelCase : Optional[int] = ( AdamW if accelerator.state.deepspeed_plugin is None or 'optimizer' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) _lowerCamelCase : Union[str, Any] = optimizer_cls(params=model.parameters() , lr=lowercase__ ) if accelerator.state.deepspeed_plugin is not None: _lowerCamelCase : str = accelerator.state.deepspeed_plugin.deepspeed_config[ 'gradient_accumulation_steps' ] else: _lowerCamelCase : Tuple = 1 _lowerCamelCase : List[Any] = (len(lowercase__ ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): _lowerCamelCase : Tuple = get_linear_schedule_with_warmup( optimizer=lowercase__ , num_warmup_steps=0 , num_training_steps=lowercase__ , ) else: _lowerCamelCase : Any = DummyScheduler(lowercase__ , total_num_steps=lowercase__ , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : Union[str, Any] = accelerator.prepare( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) # We need to keep track of how many total steps we have iterated over _lowerCamelCase : Union[str, Any] = 0 # We also need to keep track of the stating epoch so files are named properly _lowerCamelCase : Dict = 0 # Now we train the model _lowerCamelCase : Dict = evaluate.load('glue' , 'mrpc' ) _lowerCamelCase : Optional[int] = 0 _lowerCamelCase : str = {} for epoch in range(lowercase__ , lowercase__ ): model.train() for step, batch in enumerate(lowercase__ ): _lowerCamelCase : List[Any] = model(**lowercase__ ) _lowerCamelCase : int = outputs.loss _lowerCamelCase : Dict = loss / gradient_accumulation_steps accelerator.backward(lowercase__ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 model.eval() _lowerCamelCase : Union[str, Any] = 0 for step, batch in enumerate(lowercase__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _lowerCamelCase : Optional[int] = model(**lowercase__ ) _lowerCamelCase : Dict = outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times _lowerCamelCase, _lowerCamelCase : List[str] = accelerator.gather( (predictions, batch['labels']) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(lowercase__ ) - 1: _lowerCamelCase : Optional[Any] = predictions[: len(eval_dataloader.dataset ) - samples_seen] _lowerCamelCase : Dict = references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=lowercase__ , references=lowercase__ , ) _lowerCamelCase : List[Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f'''epoch {epoch}:''' , lowercase__ ) _lowerCamelCase : Tuple = eval_metric['accuracy'] if best_performance < eval_metric["accuracy"]: _lowerCamelCase : str = eval_metric['accuracy'] if args.performance_lower_bound is not None: assert ( args.performance_lower_bound <= best_performance ), f'''Best performance metric {best_performance} is lower than the lower bound {args.performance_lower_bound}''' accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , 'all_results.json' ) , 'w' ) as f: json.dump(lowercase__ , lowercase__ ) def _snake_case ( ): _lowerCamelCase : Any = argparse.ArgumentParser(description='Simple example of training script tracking peak GPU memory usage.' ) parser.add_argument( '--model_name_or_path' , type=lowercase__ , default='bert-base-cased' , help='Path to pretrained model or model identifier from huggingface.co/models.' , required=lowercase__ , ) parser.add_argument( '--output_dir' , type=lowercase__ , default='.' , help='Optional save directory where all checkpoint folders will be stored. Default is the current working directory.' , ) parser.add_argument( '--performance_lower_bound' , type=lowercase__ , default=lowercase__ , help='Optional lower bound for the performance metric. If set, the training will throw error when the performance metric drops below this value.' , ) parser.add_argument( '--num_epochs' , type=lowercase__ , default=3 , help='Number of train epochs.' , ) _lowerCamelCase : Optional[Any] = parser.parse_args() _lowerCamelCase : str = {'lr': 2E-5, 'num_epochs': args.num_epochs, 'seed': 42, 'batch_size': 16} training_function(lowercase__ , lowercase__ ) if __name__ == "__main__": main()
96
1
"""simple docstring""" from random import randint, random def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ = False , lowercase__ = False , lowercase__ = 5 , ): _lowerCamelCase : Tuple = [[-1] * number_of_cells] # Create a highway without any car _lowerCamelCase : Any = 0 _lowerCamelCase : str = max(lowercase__ , 0 ) while i < number_of_cells: _lowerCamelCase : Optional[int] = ( randint(0 , lowercase__ ) if random_speed else initial_speed ) # Place the cars i += ( randint(1 , max_speed * 2 ) if random_frequency else frequency ) # Arbitrary number, may need tuning return highway def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : Optional[int] = 0 _lowerCamelCase : Optional[int] = highway_now[car_index + 1 :] for cell in range(len(lowercase__ ) ): # May need a better name for this if cells[cell] != -1: # If the cell is not empty then return distance # we have the distance we wanted distance += 1 # Here if the car is near the end of the highway return distance + get_distance(lowercase__ , -1 ) def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : int = len(lowercase__ ) # Beforce calculations, the highway is empty _lowerCamelCase : Any = [-1] * number_of_cells for car_index in range(lowercase__ ): if highway_now[car_index] != -1: # Add 1 to the current speed of the car and cap the speed _lowerCamelCase : int = min(highway_now[car_index] + 1 , lowercase__ ) # Number of empty cell before the next car _lowerCamelCase : int = get_distance(lowercase__ , lowercase__ ) - 1 # We can't have the car causing an accident _lowerCamelCase : Union[str, Any] = min(next_highway[car_index] , lowercase__ ) if random() < probability: # Randomly, a driver will slow down _lowerCamelCase : Optional[int] = max(next_highway[car_index] - 1 , 0 ) return next_highway def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : Tuple = len(highway[0] ) for i in range(lowercase__ ): _lowerCamelCase : Dict = update(highway[i] , lowercase__ , lowercase__ ) _lowerCamelCase : Any = [-1] * number_of_cells for car_index in range(lowercase__ ): _lowerCamelCase : List[str] = next_speeds_calculated[car_index] if speed != -1: # Change the position based on the speed (with % to create the loop) _lowerCamelCase : str = (car_index + speed) % number_of_cells # Commit the change of position _lowerCamelCase : Optional[Any] = speed highway.append(lowercase__ ) return highway if __name__ == "__main__": import doctest doctest.testmod()
96
"""simple docstring""" from __future__ import annotations import copy import tempfile import unittest from transformers import CONFIG_MAPPING, AutoConfig, BertConfig, GPTaConfig, TaConfig, TapasConfig, is_tf_available from transformers.testing_utils import ( DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tensorflow_probability, require_tf, slow, ) from ..bert.test_modeling_bert import BertModelTester if is_tf_available(): from transformers import ( TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelForTableQuestionAnswering, TFAutoModelForTokenClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFFunnelBaseModel, TFFunnelModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, TFTapasForQuestionAnswering, ) from transformers.models.auto.modeling_tf_auto import ( TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_MAPPING, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.tapas.modeling_tf_tapas import TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """new-model""" if is_tf_available(): class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = NewModelConfig @require_tf class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' @slow def A_ ( self ): _lowerCamelCase : List[str] = 'bert-base-cased' _lowerCamelCase : Union[str, Any] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Union[str, Any] = TFAutoModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): _lowerCamelCase : List[str] = 'bert-base-cased' _lowerCamelCase : int = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : int = TFAutoModelForPreTraining.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Dict = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : int = TFAutoModelForCausalLM.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : str = TFAutoModelForCausalLM.from_pretrained(lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : List[Any] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : str = TFAutoModelWithLMHead.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Tuple = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Optional[int] = TFAutoModelForMaskedLM.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : Tuple = TFAutoModelForMaskedLM.from_pretrained(lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Optional[int] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Tuple = TFAutoModelForSeqaSeqLM.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : Tuple = TFAutoModelForSeqaSeqLM.from_pretrained(lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: _lowerCamelCase : str = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Union[str, Any] = TFAutoModelForSequenceClassification.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: _lowerCamelCase : Optional[Any] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : List[str] = TFAutoModelForQuestionAnswering.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow @require_tensorflow_probability def A_ ( self ): for model_name in TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST[5:6]: _lowerCamelCase : Dict = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Dict = TFAutoModelForTableQuestionAnswering.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : List[Any] = TFAutoModelForTableQuestionAnswering.from_pretrained( lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) def A_ ( self ): _lowerCamelCase : int = TFAutoModelWithLMHead.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowercase ) , 14410 ) def A_ ( self ): _lowerCamelCase : Any = TFAutoModelWithLMHead.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowercase ) , 14410 ) def A_ ( self ): # For the auto model mapping, FunnelConfig has two models: FunnelModel and FunnelBaseModel _lowerCamelCase : List[str] = TFAutoModel.from_pretrained('sgugger/funnel-random-tiny' ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Optional[int] = copy.deepcopy(model.config ) _lowerCamelCase : Dict = ['FunnelBaseModel'] _lowerCamelCase : List[Any] = TFAutoModel.from_config(lowercase ) self.assertIsInstance(lowercase , lowercase ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowercase ) _lowerCamelCase : Tuple = TFAutoModel.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) def A_ ( self ): try: AutoConfig.register('new-model' , lowercase ) _lowerCamelCase : Tuple = [ TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSequenceClassification, TFAutoModelForTokenClassification, ] for auto_class in auto_classes: with self.subTest(auto_class.__name__ ): # Wrong config class will raise an error with self.assertRaises(lowercase ): auto_class.register(lowercase , lowercase ) auto_class.register(lowercase , lowercase ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(lowercase ): auto_class.register(lowercase , lowercase ) # Now that the config is registered, it can be used as any other config with the auto-API _lowerCamelCase : Optional[Any] = BertModelTester(self ).get_config() _lowerCamelCase : Dict = NewModelConfig(**tiny_config.to_dict() ) _lowerCamelCase : int = auto_class.from_config(lowercase ) self.assertIsInstance(lowercase , lowercase ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowercase ) _lowerCamelCase : List[Any] = auto_class.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"] for mapping in ( TF_MODEL_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, ): if NewModelConfig in mapping._extra_content: del mapping._extra_content[NewModelConfig] def A_ ( self ): with self.assertRaisesRegex( lowercase , 'bert-base is not a local folder and is not a valid model identifier' ): _lowerCamelCase : Union[str, Any] = TFAutoModel.from_pretrained('bert-base' ) def A_ ( self ): with self.assertRaisesRegex( lowercase , r'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): _lowerCamelCase : str = TFAutoModel.from_pretrained(lowercase , revision='aaaaaa' ) def A_ ( self ): with self.assertRaisesRegex( lowercase , 'hf-internal-testing/config-no-model does not appear to have a file named pytorch_model.bin' , ): _lowerCamelCase : Tuple = TFAutoModel.from_pretrained('hf-internal-testing/config-no-model' ) def A_ ( self ): with self.assertRaisesRegex(lowercase , 'Use `from_pt=True` to load this model' ): _lowerCamelCase : Tuple = TFAutoModel.from_pretrained('hf-internal-testing/tiny-bert-pt-only' ) def A_ ( self ): # Make sure we have cached the model. _lowerCamelCase : Optional[int] = TFAutoModel.from_pretrained('hf-internal-testing/tiny-random-bert' ) with RequestCounter() as counter: _lowerCamelCase : Optional[int] = TFAutoModel.from_pretrained('hf-internal-testing/tiny-random-bert' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 ) # With a sharded checkpoint _lowerCamelCase : int = TFAutoModel.from_pretrained('ArthurZ/tiny-random-bert-sharded' ) with RequestCounter() as counter: _lowerCamelCase : List[Any] = TFAutoModel.from_pretrained('ArthurZ/tiny-random-bert-sharded' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 )
96
1
"""simple docstring""" import numpy as np import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModelWithProjection, PreTrainedModel from ...utils import logging lowercase__ = logging.get_logger(__name__) class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = CLIPConfig lowerCamelCase__ = ["""CLIPEncoderLayer"""] def __init__( self , lowercase ): super().__init__(lowercase ) _lowerCamelCase : List[str] = CLIPVisionModelWithProjection(config.vision_config ) _lowerCamelCase : List[Any] = nn.Linear(config.vision_config.projection_dim , 1 ) _lowerCamelCase : Dict = nn.Linear(config.vision_config.projection_dim , 1 ) @torch.no_grad() def A_ ( self , lowercase , lowercase , lowercase=0.5 , lowercase=0.5 ): _lowerCamelCase : Dict = self.vision_model(lowercase )[0] _lowerCamelCase : Any = self.p_head(lowercase ) _lowerCamelCase : str = nsfw_detected.flatten() _lowerCamelCase : int = nsfw_detected > p_threshold _lowerCamelCase : Union[str, Any] = nsfw_detected.tolist() if any(lowercase ): logger.warning( 'Potential NSFW content was detected in one or more images. A black image will be returned instead.' ' Try again with a different prompt and/or seed.' ) for idx, nsfw_detected_ in enumerate(lowercase ): if nsfw_detected_: _lowerCamelCase : Dict = np.zeros(images[idx].shape ) _lowerCamelCase : Dict = self.w_head(lowercase ) _lowerCamelCase : Tuple = watermark_detected.flatten() _lowerCamelCase : Dict = watermark_detected > w_threshold _lowerCamelCase : Dict = watermark_detected.tolist() if any(lowercase ): logger.warning( 'Potential watermarked content was detected in one or more images. A black image will be returned instead.' ' Try again with a different prompt and/or seed.' ) for idx, watermark_detected_ in enumerate(lowercase ): if watermark_detected_: _lowerCamelCase : List[str] = np.zeros(images[idx].shape ) return images, nsfw_detected, watermark_detected
96
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ = {"""configuration_ibert""": ["""IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """IBertConfig""", """IBertOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ """IBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """IBertForMaskedLM""", """IBertForMultipleChoice""", """IBertForQuestionAnswering""", """IBertForSequenceClassification""", """IBertForTokenClassification""", """IBertModel""", """IBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
96
1
"""simple docstring""" import importlib import json import os from collections import OrderedDict from typing import Dict, Optional, Union # Build the list of all feature extractors from ...configuration_utils import PretrainedConfig from ...dynamic_module_utils import get_class_from_dynamic_module, resolve_trust_remote_code from ...feature_extraction_utils import FeatureExtractionMixin from ...utils import CONFIG_NAME, FEATURE_EXTRACTOR_NAME, get_file_from_repo, logging from .auto_factory import _LazyAutoMapping from .configuration_auto import ( CONFIG_MAPPING_NAMES, AutoConfig, model_type_to_module_name, replace_list_option_in_docstrings, ) lowercase__ = logging.get_logger(__name__) lowercase__ = OrderedDict( [ ("""audio-spectrogram-transformer""", """ASTFeatureExtractor"""), ("""beit""", """BeitFeatureExtractor"""), ("""chinese_clip""", """ChineseCLIPFeatureExtractor"""), ("""clap""", """ClapFeatureExtractor"""), ("""clip""", """CLIPFeatureExtractor"""), ("""clipseg""", """ViTFeatureExtractor"""), ("""conditional_detr""", """ConditionalDetrFeatureExtractor"""), ("""convnext""", """ConvNextFeatureExtractor"""), ("""cvt""", """ConvNextFeatureExtractor"""), ("""data2vec-audio""", """Wav2Vec2FeatureExtractor"""), ("""data2vec-vision""", """BeitFeatureExtractor"""), ("""deformable_detr""", """DeformableDetrFeatureExtractor"""), ("""deit""", """DeiTFeatureExtractor"""), ("""detr""", """DetrFeatureExtractor"""), ("""dinat""", """ViTFeatureExtractor"""), ("""donut-swin""", """DonutFeatureExtractor"""), ("""dpt""", """DPTFeatureExtractor"""), ("""encodec""", """EncodecFeatureExtractor"""), ("""flava""", """FlavaFeatureExtractor"""), ("""glpn""", """GLPNFeatureExtractor"""), ("""groupvit""", """CLIPFeatureExtractor"""), ("""hubert""", """Wav2Vec2FeatureExtractor"""), ("""imagegpt""", """ImageGPTFeatureExtractor"""), ("""layoutlmv2""", """LayoutLMv2FeatureExtractor"""), ("""layoutlmv3""", """LayoutLMv3FeatureExtractor"""), ("""levit""", """LevitFeatureExtractor"""), ("""maskformer""", """MaskFormerFeatureExtractor"""), ("""mctct""", """MCTCTFeatureExtractor"""), ("""mobilenet_v1""", """MobileNetV1FeatureExtractor"""), ("""mobilenet_v2""", """MobileNetV2FeatureExtractor"""), ("""mobilevit""", """MobileViTFeatureExtractor"""), ("""nat""", """ViTFeatureExtractor"""), ("""owlvit""", """OwlViTFeatureExtractor"""), ("""perceiver""", """PerceiverFeatureExtractor"""), ("""poolformer""", """PoolFormerFeatureExtractor"""), ("""regnet""", """ConvNextFeatureExtractor"""), ("""resnet""", """ConvNextFeatureExtractor"""), ("""segformer""", """SegformerFeatureExtractor"""), ("""sew""", """Wav2Vec2FeatureExtractor"""), ("""sew-d""", """Wav2Vec2FeatureExtractor"""), ("""speech_to_text""", """Speech2TextFeatureExtractor"""), ("""speecht5""", """SpeechT5FeatureExtractor"""), ("""swiftformer""", """ViTFeatureExtractor"""), ("""swin""", """ViTFeatureExtractor"""), ("""swinv2""", """ViTFeatureExtractor"""), ("""table-transformer""", """DetrFeatureExtractor"""), ("""timesformer""", """VideoMAEFeatureExtractor"""), ("""tvlt""", """TvltFeatureExtractor"""), ("""unispeech""", """Wav2Vec2FeatureExtractor"""), ("""unispeech-sat""", """Wav2Vec2FeatureExtractor"""), ("""van""", """ConvNextFeatureExtractor"""), ("""videomae""", """VideoMAEFeatureExtractor"""), ("""vilt""", """ViltFeatureExtractor"""), ("""vit""", """ViTFeatureExtractor"""), ("""vit_mae""", """ViTFeatureExtractor"""), ("""vit_msn""", """ViTFeatureExtractor"""), ("""wav2vec2""", """Wav2Vec2FeatureExtractor"""), ("""wav2vec2-conformer""", """Wav2Vec2FeatureExtractor"""), ("""wavlm""", """Wav2Vec2FeatureExtractor"""), ("""whisper""", """WhisperFeatureExtractor"""), ("""xclip""", """CLIPFeatureExtractor"""), ("""yolos""", """YolosFeatureExtractor"""), ] ) lowercase__ = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FEATURE_EXTRACTOR_MAPPING_NAMES) def _snake_case ( lowercase__ ): for module_name, extractors in FEATURE_EXTRACTOR_MAPPING_NAMES.items(): if class_name in extractors: _lowerCamelCase : Optional[int] = model_type_to_module_name(lowercase__ ) _lowerCamelCase : List[str] = importlib.import_module(f'''.{module_name}''' , 'transformers.models' ) try: return getattr(lowercase__ , lowercase__ ) except AttributeError: continue for _, extractor in FEATURE_EXTRACTOR_MAPPING._extra_content.items(): if getattr(lowercase__ , '__name__' , lowercase__ ) == class_name: return extractor # We did not fine the class, but maybe it's because a dep is missing. In that case, the class will be in the main # init and we return the proper dummy to get an appropriate error message. _lowerCamelCase : Tuple = importlib.import_module('transformers' ) if hasattr(lowercase__ , lowercase__ ): return getattr(lowercase__ , lowercase__ ) return None def _snake_case ( lowercase__ , lowercase__ = None , lowercase__ = False , lowercase__ = False , lowercase__ = None , lowercase__ = None , lowercase__ = None , lowercase__ = False , **lowercase__ , ): _lowerCamelCase : Dict = get_file_from_repo( lowercase__ , lowercase__ , cache_dir=lowercase__ , force_download=lowercase__ , resume_download=lowercase__ , proxies=lowercase__ , use_auth_token=lowercase__ , revision=lowercase__ , local_files_only=lowercase__ , ) if resolved_config_file is None: logger.info( 'Could not locate the feature extractor configuration file, will try to use the model config instead.' ) return {} with open(lowercase__ , encoding='utf-8' ) as reader: return json.load(lowercase__ ) class lowerCAmelCase__ : '''simple docstring''' def __init__( self ): raise EnvironmentError( 'AutoFeatureExtractor is designed to be instantiated ' 'using the `AutoFeatureExtractor.from_pretrained(pretrained_model_name_or_path)` method.' ) @classmethod @replace_list_option_in_docstrings(lowercase ) def A_ ( cls , lowercase , **lowercase ): _lowerCamelCase : List[Any] = kwargs.pop('config' , lowercase ) _lowerCamelCase : int = kwargs.pop('trust_remote_code' , lowercase ) _lowerCamelCase : Any = True _lowerCamelCase, _lowerCamelCase : Tuple = FeatureExtractionMixin.get_feature_extractor_dict(lowercase , **lowercase ) _lowerCamelCase : str = config_dict.get('feature_extractor_type' , lowercase ) _lowerCamelCase : Optional[Any] = None if "AutoFeatureExtractor" in config_dict.get('auto_map' , {} ): _lowerCamelCase : Union[str, Any] = config_dict['auto_map']['AutoFeatureExtractor'] # If we don't find the feature extractor class in the feature extractor config, let's try the model config. if feature_extractor_class is None and feature_extractor_auto_map is None: if not isinstance(lowercase , lowercase ): _lowerCamelCase : List[str] = AutoConfig.from_pretrained(lowercase , **lowercase ) # It could be in `config.feature_extractor_type`` _lowerCamelCase : Dict = getattr(lowercase , 'feature_extractor_type' , lowercase ) if hasattr(lowercase , 'auto_map' ) and "AutoFeatureExtractor" in config.auto_map: _lowerCamelCase : int = config.auto_map['AutoFeatureExtractor'] if feature_extractor_class is not None: _lowerCamelCase : int = feature_extractor_class_from_name(lowercase ) _lowerCamelCase : List[str] = feature_extractor_auto_map is not None _lowerCamelCase : Union[str, Any] = feature_extractor_class is not None or type(lowercase ) in FEATURE_EXTRACTOR_MAPPING _lowerCamelCase : Dict = resolve_trust_remote_code( lowercase , lowercase , lowercase , lowercase ) if has_remote_code and trust_remote_code: _lowerCamelCase : Optional[Any] = get_class_from_dynamic_module( lowercase , lowercase , **lowercase ) _lowerCamelCase : List[Any] = kwargs.pop('code_revision' , lowercase ) if os.path.isdir(lowercase ): feature_extractor_class.register_for_auto_class() return feature_extractor_class.from_dict(lowercase , **lowercase ) elif feature_extractor_class is not None: return feature_extractor_class.from_dict(lowercase , **lowercase ) # Last try: we use the FEATURE_EXTRACTOR_MAPPING. elif type(lowercase ) in FEATURE_EXTRACTOR_MAPPING: _lowerCamelCase : Tuple = FEATURE_EXTRACTOR_MAPPING[type(lowercase )] return feature_extractor_class.from_dict(lowercase , **lowercase ) raise ValueError( F'''Unrecognized feature extractor in {pretrained_model_name_or_path}. Should have a ''' F'''`feature_extractor_type` key in its {FEATURE_EXTRACTOR_NAME} of {CONFIG_NAME}, or one of the following ''' F'''`model_type` keys in its {CONFIG_NAME}: {', '.join(c for c in FEATURE_EXTRACTOR_MAPPING_NAMES.keys() )}''' ) @staticmethod def A_ ( lowercase , lowercase ): FEATURE_EXTRACTOR_MAPPING.register(lowercase , lowercase )
96
"""simple docstring""" import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : Tuple = f'''{sampling_rate}''' _lowerCamelCase : str = '1' _lowerCamelCase : str = 'f32le' _lowerCamelCase : Union[str, Any] = [ 'ffmpeg', '-i', 'pipe:0', '-ac', ac, '-ar', ar, '-f', format_for_conversion, '-hide_banner', '-loglevel', 'quiet', 'pipe:1', ] try: with subprocess.Popen(lowercase__ , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: _lowerCamelCase : str = ffmpeg_process.communicate(lowercase__ ) except FileNotFoundError as error: raise ValueError('ffmpeg was not found but is required to load audio files from filename' ) from error _lowerCamelCase : List[Any] = output_stream[0] _lowerCamelCase : Tuple = np.frombuffer(lowercase__ , np.floataa ) if audio.shape[0] == 0: raise ValueError('Malformed soundfile' ) return audio def _snake_case ( lowercase__ , lowercase__ , lowercase__ = "f32le" , ): _lowerCamelCase : Optional[Any] = f'''{sampling_rate}''' _lowerCamelCase : List[str] = '1' if format_for_conversion == "s16le": _lowerCamelCase : List[str] = 2 elif format_for_conversion == "f32le": _lowerCamelCase : List[Any] = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) _lowerCamelCase : Dict = platform.system() if system == "Linux": _lowerCamelCase : Optional[int] = 'alsa' _lowerCamelCase : Optional[Any] = 'default' elif system == "Darwin": _lowerCamelCase : Optional[int] = 'avfoundation' _lowerCamelCase : Any = ':0' elif system == "Windows": _lowerCamelCase : Tuple = 'dshow' _lowerCamelCase : Tuple = 'default' _lowerCamelCase : Optional[int] = [ 'ffmpeg', '-f', format_, '-i', input_, '-ac', ac, '-ar', ar, '-f', format_for_conversion, '-fflags', 'nobuffer', '-hide_banner', '-loglevel', 'quiet', 'pipe:1', ] _lowerCamelCase : Tuple = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample _lowerCamelCase : List[Any] = _ffmpeg_stream(lowercase__ , lowercase__ ) for item in iterator: yield item def _snake_case ( lowercase__ , lowercase__ , lowercase__ = None , lowercase__ = None , lowercase__ = "f32le" , ): if stream_chunk_s is not None: _lowerCamelCase : int = stream_chunk_s else: _lowerCamelCase : Optional[Any] = chunk_length_s _lowerCamelCase : Optional[Any] = ffmpeg_microphone(lowercase__ , lowercase__ , format_for_conversion=lowercase__ ) if format_for_conversion == "s16le": _lowerCamelCase : List[str] = np.intaa _lowerCamelCase : str = 2 elif format_for_conversion == "f32le": _lowerCamelCase : Any = np.floataa _lowerCamelCase : List[Any] = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) if stride_length_s is None: _lowerCamelCase : Union[str, Any] = chunk_length_s / 6 _lowerCamelCase : Optional[int] = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(lowercase__ , (int, float) ): _lowerCamelCase : Any = [stride_length_s, stride_length_s] _lowerCamelCase : Tuple = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample _lowerCamelCase : Optional[Any] = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample _lowerCamelCase : List[Any] = datetime.datetime.now() _lowerCamelCase : Optional[int] = datetime.timedelta(seconds=lowercase__ ) for item in chunk_bytes_iter(lowercase__ , lowercase__ , stride=(stride_left, stride_right) , stream=lowercase__ ): # Put everything back in numpy scale _lowerCamelCase : List[Any] = np.frombuffer(item['raw'] , dtype=lowercase__ ) _lowerCamelCase : int = ( item['stride'][0] // size_of_sample, item['stride'][1] // size_of_sample, ) _lowerCamelCase : Optional[int] = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ = False ): _lowerCamelCase : int = B'' _lowerCamelCase, _lowerCamelCase : Dict = stride if stride_left + stride_right >= chunk_len: raise ValueError( f'''Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}''' ) _lowerCamelCase : str = 0 for raw in iterator: acc += raw if stream and len(lowercase__ ) < chunk_len: _lowerCamelCase : Optional[int] = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(lowercase__ ) >= chunk_len: # We are flushing the accumulator _lowerCamelCase : str = (_stride_left, stride_right) _lowerCamelCase : str = {'raw': acc[:chunk_len], 'stride': stride} if stream: _lowerCamelCase : List[Any] = False yield item _lowerCamelCase : Optional[Any] = stride_left _lowerCamelCase : str = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(lowercase__ ) > stride_left: _lowerCamelCase : Optional[Any] = {'raw': acc, 'stride': (_stride_left, 0)} if stream: _lowerCamelCase : Tuple = False yield item def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : int = 2**24 # 16Mo try: with subprocess.Popen(lowercase__ , stdout=subprocess.PIPE , bufsize=lowercase__ ) as ffmpeg_process: while True: _lowerCamelCase : Optional[Any] = ffmpeg_process.stdout.read(lowercase__ ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError('ffmpeg was not found but is required to stream audio files from filename' ) from error
96
1
"""simple docstring""" import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import ( AutoProcessor, BertTokenizerFast, BlipImageProcessor, GPTaTokenizer, InstructBlipProcessor, PreTrainedTokenizerFast, ) @require_vision class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def A_ ( self ): _lowerCamelCase : Dict = tempfile.mkdtemp() _lowerCamelCase : Union[str, Any] = BlipImageProcessor() _lowerCamelCase : Optional[int] = GPTaTokenizer.from_pretrained('hf-internal-testing/tiny-random-GPT2Model' ) _lowerCamelCase : str = BertTokenizerFast.from_pretrained('hf-internal-testing/tiny-random-bert' ) _lowerCamelCase : Optional[Any] = InstructBlipProcessor(lowercase , lowercase , lowercase ) processor.save_pretrained(self.tmpdirname ) def A_ ( self , **lowercase ): return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase ).tokenizer def A_ ( self , **lowercase ): return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase ).image_processor def A_ ( self , **lowercase ): return AutoProcessor.from_pretrained(self.tmpdirname , **lowercase ).qformer_tokenizer def A_ ( self ): shutil.rmtree(self.tmpdirname ) def A_ ( self ): _lowerCamelCase : Optional[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _lowerCamelCase : int = [Image.fromarray(np.moveaxis(lowercase , 0 , -1 ) ) for x in image_inputs] return image_inputs def A_ ( self ): _lowerCamelCase : Dict = InstructBlipProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() , qformer_tokenizer=self.get_qformer_tokenizer() , ) processor.save_pretrained(self.tmpdirname ) _lowerCamelCase : Optional[int] = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) _lowerCamelCase : str = self.get_image_processor(do_normalize=lowercase , padding_value=1.0 ) _lowerCamelCase : Union[str, Any] = InstructBlipProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=lowercase , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , lowercase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowercase ) self.assertIsInstance(processor.qformer_tokenizer , lowercase ) def A_ ( self ): _lowerCamelCase : Tuple = self.get_image_processor() _lowerCamelCase : Union[str, Any] = self.get_tokenizer() _lowerCamelCase : List[Any] = self.get_qformer_tokenizer() _lowerCamelCase : List[Any] = InstructBlipProcessor( tokenizer=lowercase , image_processor=lowercase , qformer_tokenizer=lowercase ) _lowerCamelCase : Union[str, Any] = self.prepare_image_inputs() _lowerCamelCase : Optional[Any] = image_processor(lowercase , return_tensors='np' ) _lowerCamelCase : Any = processor(images=lowercase , return_tensors='np' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def A_ ( self ): _lowerCamelCase : Tuple = self.get_image_processor() _lowerCamelCase : Tuple = self.get_tokenizer() _lowerCamelCase : Tuple = self.get_qformer_tokenizer() _lowerCamelCase : Optional[int] = InstructBlipProcessor( tokenizer=lowercase , image_processor=lowercase , qformer_tokenizer=lowercase ) _lowerCamelCase : Dict = 'lower newer' _lowerCamelCase : Dict = processor(text=lowercase ) _lowerCamelCase : Optional[Any] = tokenizer(lowercase , return_token_type_ids=lowercase ) _lowerCamelCase : Any = qformer_tokenizer(lowercase , return_token_type_ids=lowercase ) for key in encoded_tokens.keys(): self.assertListEqual(encoded_tokens[key] , encoded_processor[key] ) for key in encoded_tokens_qformer.keys(): self.assertListEqual(encoded_tokens_qformer[key] , encoded_processor['qformer_' + key] ) def A_ ( self ): _lowerCamelCase : str = self.get_image_processor() _lowerCamelCase : Optional[int] = self.get_tokenizer() _lowerCamelCase : Tuple = self.get_qformer_tokenizer() _lowerCamelCase : Union[str, Any] = InstructBlipProcessor( tokenizer=lowercase , image_processor=lowercase , qformer_tokenizer=lowercase ) _lowerCamelCase : Tuple = 'lower newer' _lowerCamelCase : Union[str, Any] = self.prepare_image_inputs() _lowerCamelCase : int = processor(text=lowercase , images=lowercase ) self.assertListEqual( list(inputs.keys() ) , ['input_ids', 'attention_mask', 'qformer_input_ids', 'qformer_attention_mask', 'pixel_values'] , ) # test if it raises when no input is passed with pytest.raises(lowercase ): processor() def A_ ( self ): _lowerCamelCase : int = self.get_image_processor() _lowerCamelCase : int = self.get_tokenizer() _lowerCamelCase : Optional[Any] = self.get_qformer_tokenizer() _lowerCamelCase : Optional[int] = InstructBlipProcessor( tokenizer=lowercase , image_processor=lowercase , qformer_tokenizer=lowercase ) _lowerCamelCase : List[str] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _lowerCamelCase : List[Any] = processor.batch_decode(lowercase ) _lowerCamelCase : str = tokenizer.batch_decode(lowercase ) self.assertListEqual(lowercase , lowercase ) def A_ ( self ): _lowerCamelCase : List[str] = self.get_image_processor() _lowerCamelCase : int = self.get_tokenizer() _lowerCamelCase : str = self.get_qformer_tokenizer() _lowerCamelCase : Any = InstructBlipProcessor( tokenizer=lowercase , image_processor=lowercase , qformer_tokenizer=lowercase ) _lowerCamelCase : int = 'lower newer' _lowerCamelCase : str = self.prepare_image_inputs() _lowerCamelCase : Union[str, Any] = processor(text=lowercase , images=lowercase ) self.assertListEqual( list(inputs.keys() ) , ['input_ids', 'attention_mask', 'qformer_input_ids', 'qformer_attention_mask', 'pixel_values'] , )
96
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = {"""ctrl""": """https://huggingface.co/ctrl/resolve/main/config.json"""} class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """ctrl""" lowerCamelCase__ = ["""past_key_values"""] lowerCamelCase__ = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowercase=246534 , lowercase=256 , lowercase=1280 , lowercase=8192 , lowercase=48 , lowercase=16 , lowercase=0.1 , lowercase=0.1 , lowercase=1E-6 , lowercase=0.02 , lowercase=True , **lowercase , ): _lowerCamelCase : Any = vocab_size _lowerCamelCase : Dict = n_positions _lowerCamelCase : Optional[int] = n_embd _lowerCamelCase : str = n_layer _lowerCamelCase : Union[str, Any] = n_head _lowerCamelCase : Any = dff _lowerCamelCase : int = resid_pdrop _lowerCamelCase : Dict = embd_pdrop _lowerCamelCase : Union[str, Any] = layer_norm_epsilon _lowerCamelCase : Tuple = initializer_range _lowerCamelCase : str = use_cache super().__init__(**lowercase )
96
1
"""simple docstring""" lowercase__ = range(2, 20 + 1) lowercase__ = [10**k for k in range(ks[-1] + 1)] lowercase__ = {} def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : Optional[Any] = sum(a_i[j] for j in range(lowercase__ , len(lowercase__ ) ) ) _lowerCamelCase : Dict = sum(a_i[j] * base[j] for j in range(min(len(lowercase__ ) , lowercase__ ) ) ) _lowerCamelCase, _lowerCamelCase : Dict = 0, 0 _lowerCamelCase : List[str] = n - i _lowerCamelCase : Optional[int] = memo.get(lowercase__ ) if sub_memo is not None: _lowerCamelCase : Tuple = sub_memo.get(lowercase__ ) if jumps is not None and len(lowercase__ ) > 0: # find and make the largest jump without going over _lowerCamelCase : Optional[Any] = -1 for _k in range(len(lowercase__ ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: _lowerCamelCase : Dict = _k break if max_jump >= 0: _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : Dict = jumps[max_jump] # since the difference between jumps is cached, add c _lowerCamelCase : Any = diff + c for j in range(min(lowercase__ , len(lowercase__ ) ) ): _lowerCamelCase, _lowerCamelCase : Tuple = divmod(lowercase__ , 10 ) if new_c > 0: add(lowercase__ , lowercase__ , lowercase__ ) else: _lowerCamelCase : List[Any] = [] else: _lowerCamelCase : List[str] = {c: []} _lowerCamelCase : str = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps _lowerCamelCase, _lowerCamelCase : Any = next_term(lowercase__ , k - 1 , i + dn , lowercase__ ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead _lowerCamelCase, _lowerCamelCase : str = compute(lowercase__ , lowercase__ , i + dn , lowercase__ ) diff += _diff dn += terms_jumped _lowerCamelCase : Union[str, Any] = sub_memo[c] # keep jumps sorted by # of terms skipped _lowerCamelCase : Dict = 0 while j < len(lowercase__ ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(lowercase__ , (diff, dn, k) ) return (diff, dn) def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): if i >= n: return 0, i if k > len(lowercase__ ): a_i.extend([0 for _ in range(k - len(lowercase__ ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) _lowerCamelCase : int = i _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : int = 0, 0, 0 for j in range(len(lowercase__ ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 _lowerCamelCase : int = ds_c + ds_b diff += addend _lowerCamelCase : Optional[int] = 0 for j in range(lowercase__ ): _lowerCamelCase : Dict = a_i[j] + addend _lowerCamelCase, _lowerCamelCase : List[Any] = divmod(lowercase__ , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(lowercase__ , lowercase__ , lowercase__ ) return diff, i - start_i def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): for j in range(lowercase__ , len(lowercase__ ) ): _lowerCamelCase : int = digits[j] + addend if s >= 10: _lowerCamelCase, _lowerCamelCase : Dict = divmod(lowercase__ , 10 ) _lowerCamelCase : Union[str, Any] = addend // 10 + quotient else: _lowerCamelCase : Dict = s _lowerCamelCase : Optional[Any] = addend // 10 if addend == 0: break while addend > 0: _lowerCamelCase, _lowerCamelCase : Dict = divmod(lowercase__ , 10 ) digits.append(lowercase__ ) def _snake_case ( lowercase__ = 10**15 ): _lowerCamelCase : Optional[Any] = [1] _lowerCamelCase : int = 1 _lowerCamelCase : Any = 0 while True: _lowerCamelCase, _lowerCamelCase : Dict = next_term(lowercase__ , 20 , i + dn , lowercase__ ) dn += terms_jumped if dn == n - i: break _lowerCamelCase : int = 0 for j in range(len(lowercase__ ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(F"{solution() = }")
96
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator from typing import Any class lowerCAmelCase__ : '''simple docstring''' def __init__( self , lowercase ): _lowerCamelCase : Any = data _lowerCamelCase : Node | None = None class lowerCAmelCase__ : '''simple docstring''' def __init__( self ): _lowerCamelCase : str = None _lowerCamelCase : str = None def __iter__( self ): _lowerCamelCase : List[str] = self.head while self.head: yield node.data _lowerCamelCase : Optional[int] = node.next if node == self.head: break def __len__( self ): return sum(1 for _ in self ) def __repr__( self ): return "->".join(str(lowercase ) for item in iter(self ) ) def A_ ( self , lowercase ): self.insert_nth(len(self ) , lowercase ) def A_ ( self , lowercase ): self.insert_nth(0 , lowercase ) def A_ ( self , lowercase , lowercase ): if index < 0 or index > len(self ): raise IndexError('list index out of range.' ) _lowerCamelCase : List[Any] = Node(lowercase ) if self.head is None: _lowerCamelCase : str = new_node # first node points itself _lowerCamelCase : Union[str, Any] = new_node elif index == 0: # insert at head _lowerCamelCase : List[str] = self.head _lowerCamelCase : str = new_node else: _lowerCamelCase : Union[str, Any] = self.head for _ in range(index - 1 ): _lowerCamelCase : List[Any] = temp.next _lowerCamelCase : Union[str, Any] = temp.next _lowerCamelCase : List[str] = new_node if index == len(self ) - 1: # insert at tail _lowerCamelCase : Any = new_node def A_ ( self ): return self.delete_nth(0 ) def A_ ( self ): return self.delete_nth(len(self ) - 1 ) def A_ ( self , lowercase = 0 ): if not 0 <= index < len(self ): raise IndexError('list index out of range.' ) _lowerCamelCase : Any = self.head if self.head == self.tail: # just one node _lowerCamelCase : List[str] = None elif index == 0: # delete head node _lowerCamelCase : List[str] = self.tail.next.next _lowerCamelCase : Optional[int] = self.head.next else: _lowerCamelCase : Dict = self.head for _ in range(index - 1 ): _lowerCamelCase : List[Any] = temp.next _lowerCamelCase : int = temp.next _lowerCamelCase : Optional[int] = temp.next.next if index == len(self ) - 1: # delete at tail _lowerCamelCase : List[Any] = temp return delete_node.data def A_ ( self ): return len(self ) == 0 def _snake_case ( ): _lowerCamelCase : Union[str, Any] = CircularLinkedList() assert len(lowercase__ ) == 0 assert circular_linked_list.is_empty() is True assert str(lowercase__ ) == "" try: circular_linked_list.delete_front() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_tail() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_nth(-1 ) raise AssertionError except IndexError: assert True try: circular_linked_list.delete_nth(0 ) raise AssertionError except IndexError: assert True assert circular_linked_list.is_empty() is True for i in range(5 ): assert len(lowercase__ ) == i circular_linked_list.insert_nth(lowercase__ , i + 1 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 6 ) ) circular_linked_list.insert_tail(6 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 7 ) ) circular_linked_list.insert_head(0 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(0 , 7 ) ) assert circular_linked_list.delete_front() == 0 assert circular_linked_list.delete_tail() == 6 assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 6 ) ) assert circular_linked_list.delete_nth(2 ) == 3 circular_linked_list.insert_nth(2 , 3 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 6 ) ) assert circular_linked_list.is_empty() is False if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" import math from typing import Callable, List, Optional, Union import numpy as np import PIL import torch from PIL import Image from transformers import CLIPTextModel, CLIPTokenizer from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from diffusers.schedulers import DDIMScheduler, DDPMScheduler, LMSDiscreteScheduler, PNDMScheduler def _snake_case ( lowercase__ , lowercase__ , lowercase__=[] ): _lowerCamelCase : Dict = size[0] - overlap_pixels * 2 _lowerCamelCase : int = size[1] - overlap_pixels * 2 for letter in ["l", "r"]: if letter in remove_borders: size_x += overlap_pixels for letter in ["t", "b"]: if letter in remove_borders: size_y += overlap_pixels _lowerCamelCase : Any = np.ones((size_y, size_x) , dtype=np.uinta ) * 255 _lowerCamelCase : List[str] = np.pad(lowercase__ , mode='linear_ramp' , pad_width=lowercase__ , end_values=0 ) if "l" in remove_borders: _lowerCamelCase : Tuple = mask[:, overlap_pixels : mask.shape[1]] if "r" in remove_borders: _lowerCamelCase : Any = mask[:, 0 : mask.shape[1] - overlap_pixels] if "t" in remove_borders: _lowerCamelCase : Optional[Any] = mask[overlap_pixels : mask.shape[0], :] if "b" in remove_borders: _lowerCamelCase : Union[str, Any] = mask[0 : mask.shape[0] - overlap_pixels, :] return mask def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): return max(lowercase__ , min(lowercase__ , lowercase__ ) ) def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): return ( clamp(rect[0] , min[0] , max[0] ), clamp(rect[1] , min[1] , max[1] ), clamp(rect[2] , min[0] , max[0] ), clamp(rect[3] , min[1] , max[1] ), ) def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : Any = list(lowercase__ ) rect[0] -= overlap rect[1] -= overlap rect[2] += overlap rect[3] += overlap _lowerCamelCase : str = clamp_rect(lowercase__ , [0, 0] , [image_size[0], image_size[1]] ) return rect def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : Tuple = Image.new('RGB' , (tile.size[0] + original_slice, tile.size[1]) ) result.paste( original_image.resize((tile.size[0], tile.size[1]) , Image.BICUBIC ).crop( (slice_x, 0, slice_x + original_slice, tile.size[1]) ) , (0, 0) , ) result.paste(lowercase__ , (original_slice, 0) ) return result def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : Optional[int] = (original_image_slice * 4, 0, tile.size[0], tile.size[1]) _lowerCamelCase : Optional[int] = tile.crop(lowercase__ ) return tile def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : Tuple = n % d return n - divisor class lowerCAmelCase__ ( lowercase ): '''simple docstring''' def __init__( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase = 350 , ): super().__init__( vae=lowercase , text_encoder=lowercase , tokenizer=lowercase , unet=lowercase , low_res_scheduler=lowercase , scheduler=lowercase , max_noise_level=lowercase , ) def A_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , **lowercase ): torch.manual_seed(0 ) _lowerCamelCase : Optional[int] = ( min(image.size[0] - (tile_size + original_image_slice) , x * tile_size ), min(image.size[1] - (tile_size + original_image_slice) , y * tile_size ), min(image.size[0] , (x + 1) * tile_size ), min(image.size[1] , (y + 1) * tile_size ), ) _lowerCamelCase : str = add_overlap_rect(lowercase , lowercase , image.size ) _lowerCamelCase : Dict = image.crop(lowercase ) _lowerCamelCase : Optional[int] = ((crop_rect[0] + ((crop_rect[2] - crop_rect[0]) / 2)) / image.size[0]) * tile.size[0] _lowerCamelCase : List[str] = translated_slice_x - (original_image_slice / 2) _lowerCamelCase : List[Any] = max(0 , lowercase ) _lowerCamelCase : Optional[Any] = squeeze_tile(lowercase , lowercase , lowercase , lowercase ) _lowerCamelCase : int = to_input.size _lowerCamelCase : Union[str, Any] = to_input.resize((tile_size, tile_size) , Image.BICUBIC ) _lowerCamelCase : Union[str, Any] = super(lowercase , self ).__call__(image=lowercase , **lowercase ).images[0] _lowerCamelCase : Tuple = upscaled_tile.resize((orig_input_size[0] * 4, orig_input_size[1] * 4) , Image.BICUBIC ) _lowerCamelCase : List[Any] = unsqueeze_tile(lowercase , lowercase ) _lowerCamelCase : Dict = upscaled_tile.resize((tile.size[0] * 4, tile.size[1] * 4) , Image.BICUBIC ) _lowerCamelCase : Dict = [] if x == 0: remove_borders.append('l' ) elif crop_rect[2] == image.size[0]: remove_borders.append('r' ) if y == 0: remove_borders.append('t' ) elif crop_rect[3] == image.size[1]: remove_borders.append('b' ) _lowerCamelCase : str = Image.fromarray( make_transparency_mask( (upscaled_tile.size[0], upscaled_tile.size[1]) , tile_border * 4 , remove_borders=lowercase ) , mode='L' , ) final_image.paste( lowercase , (crop_rect_with_overlap[0] * 4, crop_rect_with_overlap[1] * 4) , lowercase ) @torch.no_grad() def __call__( self , lowercase , lowercase , lowercase = 75 , lowercase = 9.0 , lowercase = 50 , lowercase = None , lowercase = 1 , lowercase = 0.0 , lowercase = None , lowercase = None , lowercase = None , lowercase = 1 , lowercase = 128 , lowercase = 32 , lowercase = 32 , ): _lowerCamelCase : Any = Image.new('RGB' , (image.size[0] * 4, image.size[1] * 4) ) _lowerCamelCase : Optional[int] = math.ceil(image.size[0] / tile_size ) _lowerCamelCase : Optional[Any] = math.ceil(image.size[1] / tile_size ) _lowerCamelCase : Dict = tcx * tcy _lowerCamelCase : List[str] = 0 for y in range(lowercase ): for x in range(lowercase ): self._process_tile( lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , prompt=lowercase , num_inference_steps=lowercase , guidance_scale=lowercase , noise_level=lowercase , negative_prompt=lowercase , num_images_per_prompt=lowercase , eta=lowercase , generator=lowercase , latents=lowercase , ) current_count += 1 if callback is not None: callback({'progress': current_count / total_tile_count, 'image': final_image} ) return final_image def _snake_case ( ): # Run a demo _lowerCamelCase : int = 'stabilityai/stable-diffusion-x4-upscaler' _lowerCamelCase : Optional[Any] = StableDiffusionTiledUpscalePipeline.from_pretrained(lowercase__ , revision='fp16' , torch_dtype=torch.floataa ) _lowerCamelCase : Optional[int] = pipe.to('cuda' ) _lowerCamelCase : List[Any] = Image.open('../../docs/source/imgs/diffusers_library.jpg' ) def callback(lowercase__ ): print(f'''progress: {obj['progress']:.4f}''' ) obj["image"].save('diffusers_library_progress.jpg' ) _lowerCamelCase : str = pipe(image=lowercase__ , prompt='Black font, white background, vector' , noise_level=40 , callback=lowercase__ ) final_image.save('diffusers_library.jpg' ) if __name__ == "__main__": main()
96
"""simple docstring""" # Lint as: python3 import os import re import urllib.parse from pathlib import Path from typing import Callable, List, Optional, Union from zipfile import ZipFile from ..utils.file_utils import cached_path, hf_github_url from ..utils.logging import get_logger from ..utils.version import Version lowercase__ = get_logger(__name__) class lowerCAmelCase__ : '''simple docstring''' lowerCamelCase__ = """dummy_data""" lowerCamelCase__ = """datasets""" lowerCamelCase__ = False def __init__( self , lowercase , lowercase , lowercase , lowercase = None , lowercase = False , lowercase = True , lowercase = None , ): _lowerCamelCase : Optional[Any] = 0 _lowerCamelCase : Dict = dataset_name _lowerCamelCase : Union[str, Any] = cache_dir _lowerCamelCase : Dict = use_local_dummy_data _lowerCamelCase : Tuple = config # download_callbacks take a single url as input _lowerCamelCase : List[Callable] = download_callbacks or [] # if False, it doesn't load existing files and it returns the paths of the dummy files relative # to the dummy_data zip file root _lowerCamelCase : Any = load_existing_dummy_data # TODO(PVP, QL) might need to make this more general _lowerCamelCase : str = str(lowercase ) # to be downloaded _lowerCamelCase : Union[str, Any] = None _lowerCamelCase : int = None @property def A_ ( self ): if self._dummy_file is None: _lowerCamelCase : Tuple = self.download_dummy_data() return self._dummy_file @property def A_ ( self ): if self.config is not None: # structure is dummy / config_name / version_name return os.path.join('dummy' , self.config.name , self.version_name ) # structure is dummy / version_name return os.path.join('dummy' , self.version_name ) @property def A_ ( self ): return os.path.join(self.dummy_data_folder , 'dummy_data.zip' ) def A_ ( self ): _lowerCamelCase : List[str] = ( self.local_path_to_dummy_data if self.use_local_dummy_data is True else self.github_path_to_dummy_data ) _lowerCamelCase : int = cached_path( lowercase , cache_dir=self.cache_dir , extract_compressed_file=lowercase , force_extract=lowercase ) return os.path.join(lowercase , self.dummy_file_name ) @property def A_ ( self ): return os.path.join(self.datasets_scripts_dir , self.dataset_name , self.dummy_zip_file ) @property def A_ ( self ): if self._bucket_url is None: _lowerCamelCase : List[Any] = hf_github_url(self.dataset_name , self.dummy_zip_file.replace(os.sep , '/' ) ) return self._bucket_url @property def A_ ( self ): # return full path if its a dir if os.path.isdir(self.dummy_file ): return self.dummy_file # else cut off path to file -> example `xsum`. return "/".join(self.dummy_file.replace(os.sep , '/' ).split('/' )[:-1] ) def A_ ( self , lowercase , *lowercase ): if self.load_existing_dummy_data: # dummy data is downloaded and tested _lowerCamelCase : Union[str, Any] = self.dummy_file else: # dummy data cannot be downloaded and only the path to dummy file is returned _lowerCamelCase : Union[str, Any] = self.dummy_file_name # special case when data_url is a dict if isinstance(lowercase , lowercase ): return self.create_dummy_data_dict(lowercase , lowercase ) elif isinstance(lowercase , (list, tuple) ): return self.create_dummy_data_list(lowercase , lowercase ) else: return self.create_dummy_data_single(lowercase , lowercase ) def A_ ( self , lowercase , *lowercase ): return self.download_and_extract(lowercase ) def A_ ( self , lowercase , lowercase ): return self.download_and_extract(lowercase ) def A_ ( self , lowercase , *lowercase , **lowercase ): return path def A_ ( self ): return {} def A_ ( self , lowercase , lowercase ): _lowerCamelCase : Optional[int] = {} for key, single_urls in data_url.items(): for download_callback in self.download_callbacks: if isinstance(lowercase , lowercase ): for single_url in single_urls: download_callback(lowercase ) else: _lowerCamelCase : List[Any] = single_urls download_callback(lowercase ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus if isinstance(lowercase , lowercase ): _lowerCamelCase : List[Any] = [os.path.join(lowercase , urllib.parse.quote_plus(Path(lowercase ).name ) ) for x in single_urls] else: _lowerCamelCase : Optional[int] = single_urls _lowerCamelCase : List[Any] = os.path.join(lowercase , urllib.parse.quote_plus(Path(lowercase ).name ) ) _lowerCamelCase : int = value # make sure that values are unique if all(isinstance(lowercase , lowercase ) for i in dummy_data_dict.values() ) and len(set(dummy_data_dict.values() ) ) < len( dummy_data_dict.values() ): # append key to value to make its name unique _lowerCamelCase : List[Any] = {key: value + key for key, value in dummy_data_dict.items()} return dummy_data_dict def A_ ( self , lowercase , lowercase ): _lowerCamelCase : Optional[Any] = [] # trick: if there are many shards named like `data.txt-000001-of-00300`, only use the first one _lowerCamelCase : List[str] = all(bool(re.findall('[0-9]{3,}-of-[0-9]{3,}' , lowercase ) ) for url in data_url ) _lowerCamelCase : int = all( url.startswith('https://ftp.ncbi.nlm.nih.gov/pubmed/baseline/pubmed' ) for url in data_url ) if data_url and (is_tf_records or is_pubmed_records): _lowerCamelCase : List[str] = [data_url[0]] * len(lowercase ) for single_url in data_url: for download_callback in self.download_callbacks: download_callback(lowercase ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus _lowerCamelCase : str = os.path.join(lowercase , urllib.parse.quote_plus(single_url.split('/' )[-1] ) ) dummy_data_list.append(lowercase ) return dummy_data_list def A_ ( self , lowercase , lowercase ): for download_callback in self.download_callbacks: download_callback(lowercase ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus _lowerCamelCase : Tuple = os.path.join(lowercase , urllib.parse.quote_plus(data_url.split('/' )[-1] ) ) if os.path.exists(lowercase ) or not self.load_existing_dummy_data: return value else: # Backward compatibility, maybe deprecate at one point. # For many datasets with single url calls to dl_manager.download_and_extract, # the dummy_data.zip file is actually the zipped downloaded file # while now we expected the dummy_data.zip file to be a directory containing # the downloaded file. return path_to_dummy_data def A_ ( self ): pass def A_ ( self ): pass def A_ ( self , lowercase ): def _iter_archive_members(lowercase ): # this preserves the order of the members inside the ZIP archive _lowerCamelCase : str = Path(self.dummy_file ).parent _lowerCamelCase : Union[str, Any] = path.relative_to(lowercase ) with ZipFile(self.local_path_to_dummy_data ) as zip_file: _lowerCamelCase : List[str] = zip_file.namelist() for member in members: if member.startswith(relative_path.as_posix() ): yield dummy_parent_path.joinpath(lowercase ) _lowerCamelCase : Optional[int] = Path(lowercase ) _lowerCamelCase : Dict = _iter_archive_members(lowercase ) if self.use_local_dummy_data else path.rglob('*' ) for file_path in file_paths: if file_path.is_file() and not file_path.name.startswith(('.', '__') ): yield file_path.relative_to(lowercase ).as_posix(), file_path.open('rb' ) def A_ ( self , lowercase ): if not isinstance(lowercase , lowercase ): _lowerCamelCase : List[str] = [paths] for path in paths: if os.path.isfile(lowercase ): if os.path.basename(lowercase ).startswith(('.', '__') ): return yield path else: for dirpath, dirnames, filenames in os.walk(lowercase ): if os.path.basename(lowercase ).startswith(('.', '__') ): continue dirnames.sort() for filename in sorted(lowercase ): if filename.startswith(('.', '__') ): continue yield os.path.join(lowercase , lowercase )
96
1
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import YolosConfig, YolosForObjectDetection, YolosImageProcessor from transformers.utils import logging logging.set_verbosity_info() lowercase__ = logging.get_logger(__name__) def _snake_case ( lowercase__ ): _lowerCamelCase : Tuple = YolosConfig() # size of the architecture if "yolos_ti" in yolos_name: _lowerCamelCase : Tuple = 192 _lowerCamelCase : List[Any] = 768 _lowerCamelCase : Optional[Any] = 12 _lowerCamelCase : Dict = 3 _lowerCamelCase : Optional[Any] = [800, 1333] _lowerCamelCase : List[Any] = False elif yolos_name == "yolos_s_dWr": _lowerCamelCase : List[str] = 330 _lowerCamelCase : List[str] = 14 _lowerCamelCase : List[str] = 6 _lowerCamelCase : List[Any] = 1320 elif "yolos_s" in yolos_name: _lowerCamelCase : int = 384 _lowerCamelCase : Tuple = 1536 _lowerCamelCase : Dict = 12 _lowerCamelCase : List[Any] = 6 elif "yolos_b" in yolos_name: _lowerCamelCase : Any = [800, 1344] _lowerCamelCase : str = 91 _lowerCamelCase : Tuple = 'huggingface/label-files' _lowerCamelCase : Optional[Any] = 'coco-detection-id2label.json' _lowerCamelCase : Tuple = json.load(open(hf_hub_download(lowercase__ , lowercase__ , repo_type='dataset' ) , 'r' ) ) _lowerCamelCase : Tuple = {int(lowercase__ ): v for k, v in idalabel.items()} _lowerCamelCase : List[Any] = idalabel _lowerCamelCase : str = {v: k for k, v in idalabel.items()} return config def _snake_case ( lowercase__ , lowercase__ , lowercase__ = False ): for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _lowerCamelCase : List[str] = state_dict.pop(f'''blocks.{i}.attn.qkv.weight''' ) _lowerCamelCase : Dict = state_dict.pop(f'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict _lowerCamelCase : Tuple = in_proj_weight[: config.hidden_size, :] _lowerCamelCase : Tuple = in_proj_bias[: config.hidden_size] _lowerCamelCase : List[str] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _lowerCamelCase : List[Any] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _lowerCamelCase : Optional[Any] = in_proj_weight[-config.hidden_size :, :] _lowerCamelCase : List[str] = in_proj_bias[-config.hidden_size :] def _snake_case ( lowercase__ ): if "backbone" in name: _lowerCamelCase : Any = name.replace('backbone' , 'vit' ) if "cls_token" in name: _lowerCamelCase : int = name.replace('cls_token' , 'embeddings.cls_token' ) if "det_token" in name: _lowerCamelCase : int = name.replace('det_token' , 'embeddings.detection_tokens' ) if "mid_pos_embed" in name: _lowerCamelCase : Any = name.replace('mid_pos_embed' , 'encoder.mid_position_embeddings' ) if "pos_embed" in name: _lowerCamelCase : Optional[int] = name.replace('pos_embed' , 'embeddings.position_embeddings' ) if "patch_embed.proj" in name: _lowerCamelCase : str = name.replace('patch_embed.proj' , 'embeddings.patch_embeddings.projection' ) if "blocks" in name: _lowerCamelCase : Union[str, Any] = name.replace('blocks' , 'encoder.layer' ) if "attn.proj" in name: _lowerCamelCase : str = name.replace('attn.proj' , 'attention.output.dense' ) if "attn" in name: _lowerCamelCase : List[Any] = name.replace('attn' , 'attention.self' ) if "norm1" in name: _lowerCamelCase : Union[str, Any] = name.replace('norm1' , 'layernorm_before' ) if "norm2" in name: _lowerCamelCase : List[str] = name.replace('norm2' , 'layernorm_after' ) if "mlp.fc1" in name: _lowerCamelCase : Union[str, Any] = name.replace('mlp.fc1' , 'intermediate.dense' ) if "mlp.fc2" in name: _lowerCamelCase : Optional[Any] = name.replace('mlp.fc2' , 'output.dense' ) if "class_embed" in name: _lowerCamelCase : List[str] = name.replace('class_embed' , 'class_labels_classifier' ) if "bbox_embed" in name: _lowerCamelCase : Dict = name.replace('bbox_embed' , 'bbox_predictor' ) if "vit.norm" in name: _lowerCamelCase : List[Any] = name.replace('vit.norm' , 'vit.layernorm' ) return name def _snake_case ( lowercase__ , lowercase__ ): for key in orig_state_dict.copy().keys(): _lowerCamelCase : Dict = orig_state_dict.pop(lowercase__ ) if "qkv" in key: _lowerCamelCase : Optional[Any] = key.split('.' ) _lowerCamelCase : Union[str, Any] = int(key_split[2] ) _lowerCamelCase : int = model.vit.encoder.layer[layer_num].attention.attention.all_head_size if "weight" in key: _lowerCamelCase : Optional[Any] = val[:dim, :] _lowerCamelCase : Any = val[ dim : dim * 2, : ] _lowerCamelCase : Union[str, Any] = val[-dim:, :] else: _lowerCamelCase : Dict = val[:dim] _lowerCamelCase : int = val[dim : dim * 2] _lowerCamelCase : List[Any] = val[-dim:] else: _lowerCamelCase : List[str] = val return orig_state_dict def _snake_case ( ): _lowerCamelCase : Union[str, Any] = 'http://images.cocodataset.org/val2017/000000039769.jpg' _lowerCamelCase : Dict = Image.open(requests.get(lowercase__ , stream=lowercase__ ).raw ) return im @torch.no_grad() def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ = False ): _lowerCamelCase : str = get_yolos_config(lowercase__ ) # load original state_dict _lowerCamelCase : Union[str, Any] = torch.load(lowercase__ , map_location='cpu' )['model'] # load 🤗 model _lowerCamelCase : List[Any] = YolosForObjectDetection(lowercase__ ) model.eval() _lowerCamelCase : str = convert_state_dict(lowercase__ , lowercase__ ) model.load_state_dict(lowercase__ ) # Check outputs on an image, prepared by YolosImageProcessor _lowerCamelCase : Optional[int] = 800 if yolos_name != 'yolos_ti' else 512 _lowerCamelCase : int = YolosImageProcessor(format='coco_detection' , size=lowercase__ ) _lowerCamelCase : int = image_processor(images=prepare_img() , return_tensors='pt' ) _lowerCamelCase : int = model(**lowercase__ ) _lowerCamelCase, _lowerCamelCase : Any = outputs.logits, outputs.pred_boxes _lowerCamelCase, _lowerCamelCase : str = None, None if yolos_name == "yolos_ti": _lowerCamelCase : List[Any] = torch.tensor( [[-3_9.5_0_2_2, -1_1.9_8_2_0, -1_7.6_8_8_8], [-2_9.9_5_7_4, -9.9_7_6_9, -1_7.7_6_9_1], [-4_2.3_2_8_1, -2_0.7_2_0_0, -3_0.6_2_9_4]] ) _lowerCamelCase : int = torch.tensor( [[0.4_0_2_1, 0.0_8_3_6, 0.7_9_7_9], [0.0_1_8_4, 0.2_6_0_9, 0.0_3_6_4], [0.1_7_8_1, 0.2_0_0_4, 0.2_0_9_5]] ) elif yolos_name == "yolos_s_200_pre": _lowerCamelCase : Optional[int] = torch.tensor( [[-2_4.0_2_4_8, -1_0.3_0_2_4, -1_4.8_2_9_0], [-4_2.0_3_9_2, -1_6.8_2_0_0, -2_7.4_3_3_4], [-2_7.2_7_4_3, -1_1.8_1_5_4, -1_8.7_1_4_8]] ) _lowerCamelCase : Union[str, Any] = torch.tensor( [[0.2_5_5_9, 0.5_4_5_5, 0.4_7_0_6], [0.2_9_8_9, 0.7_2_7_9, 0.1_8_7_5], [0.7_7_3_2, 0.4_0_1_7, 0.4_4_6_2]] ) elif yolos_name == "yolos_s_300_pre": _lowerCamelCase : Optional[int] = torch.tensor( [[-3_6.2_2_2_0, -1_4.4_3_8_5, -2_3.5_4_5_7], [-3_5.6_9_7_0, -1_4.7_5_8_3, -2_1.3_9_3_5], [-3_1.5_9_3_9, -1_3.6_0_4_2, -1_6.8_0_4_9]] ) _lowerCamelCase : List[str] = torch.tensor( [[0.7_6_1_4, 0.2_3_1_6, 0.4_7_2_8], [0.7_1_6_8, 0.4_4_9_5, 0.3_8_5_5], [0.4_9_9_6, 0.1_4_6_6, 0.9_9_9_6]] ) elif yolos_name == "yolos_s_dWr": _lowerCamelCase : Dict = torch.tensor( [[-4_2.8_6_6_8, -2_4.1_0_4_9, -4_1.1_6_9_0], [-3_4.7_4_5_6, -1_4.1_2_7_4, -2_4.9_1_9_4], [-3_3.7_8_9_8, -1_2.1_9_4_6, -2_5.6_4_9_5]] ) _lowerCamelCase : List[Any] = torch.tensor( [[0.5_5_8_7, 0.2_7_7_3, 0.0_6_0_5], [0.5_0_0_4, 0.3_0_1_4, 0.9_9_9_4], [0.4_9_9_9, 0.1_5_4_8, 0.9_9_9_4]] ) elif yolos_name == "yolos_base": _lowerCamelCase : Optional[Any] = torch.tensor( [[-4_0.6_0_6_4, -2_4.3_0_8_4, -3_2.6_4_4_7], [-5_5.1_9_9_0, -3_0.7_7_1_9, -3_5.5_8_7_7], [-5_1.4_3_1_1, -3_3.3_5_0_7, -3_5.6_4_6_2]] ) _lowerCamelCase : Union[str, Any] = torch.tensor( [[0.5_5_5_5, 0.2_7_9_4, 0.0_6_5_5], [0.9_0_4_9, 0.2_6_6_4, 0.1_8_9_4], [0.9_1_8_3, 0.1_9_8_4, 0.1_6_3_5]] ) else: raise ValueError(f'''Unknown yolos_name: {yolos_name}''' ) assert torch.allclose(logits[0, :3, :3] , lowercase__ , atol=1E-4 ) assert torch.allclose(pred_boxes[0, :3, :3] , lowercase__ , atol=1E-4 ) Path(lowercase__ ).mkdir(exist_ok=lowercase__ ) print(f'''Saving model {yolos_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowercase__ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(lowercase__ ) if push_to_hub: _lowerCamelCase : int = { 'yolos_ti': 'yolos-tiny', 'yolos_s_200_pre': 'yolos-small', 'yolos_s_300_pre': 'yolos-small-300', 'yolos_s_dWr': 'yolos-small-dwr', 'yolos_base': 'yolos-base', } print('Pushing to the hub...' ) _lowerCamelCase : List[Any] = model_mapping[yolos_name] image_processor.push_to_hub(lowercase__ , organization='hustvl' ) model.push_to_hub(lowercase__ , organization='hustvl' ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--yolos_name""", default="""yolos_s_200_pre""", type=str, help=( """Name of the YOLOS model you'd like to convert. Should be one of 'yolos_ti', 'yolos_s_200_pre',""" """ 'yolos_s_300_pre', 'yolos_s_dWr', 'yolos_base'.""" ), ) parser.add_argument( """--checkpoint_path""", default=None, type=str, help="""Path to the original state dict (.pth file).""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) lowercase__ = parser.parse_args() convert_yolos_checkpoint(args.yolos_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
96
"""simple docstring""" def _snake_case ( lowercase__ ): stooge(lowercase__ , 0 , len(lowercase__ ) - 1 ) return arr def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: _lowerCamelCase, _lowerCamelCase : Optional[Any] = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: _lowerCamelCase : Union[str, Any] = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(lowercase__ , lowercase__ , (h - t) ) # Recursively sort last 2/3 elements stooge(lowercase__ , i + t , (lowercase__) ) # Recursively sort first 2/3 elements stooge(lowercase__ , lowercase__ , (h - t) ) if __name__ == "__main__": lowercase__ = input("""Enter numbers separated by a comma:\n""").strip() lowercase__ = [int(item) for item in user_input.split(""",""")] print(stooge_sort(unsorted))
96
1
"""simple docstring""" def _snake_case ( lowercase__ , lowercase__ ): # "extended trapezoidal rule" # int(f) = dx/2 * (f1 + 2f2 + ... + fn) _lowerCamelCase : List[Any] = (boundary[1] - boundary[0]) / steps _lowerCamelCase : Tuple = boundary[0] _lowerCamelCase : Dict = boundary[1] _lowerCamelCase : List[Any] = make_points(lowercase__ , lowercase__ , lowercase__ ) _lowerCamelCase : List[Any] = 0.0 y += (h / 2.0) * f(lowercase__ ) for i in x_i: # print(i) y += h * f(lowercase__ ) y += (h / 2.0) * f(lowercase__ ) return y def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : str = a + h while x < (b - h): yield x _lowerCamelCase : int = x + h def _snake_case ( lowercase__ ): # enter your function here _lowerCamelCase : Optional[Any] = (x - 0) * (x - 0) return y def _snake_case ( ): _lowerCamelCase : int = 0.0 # Lower bound of integration _lowerCamelCase : Optional[int] = 1.0 # Upper bound of integration _lowerCamelCase : List[str] = 1_0.0 # define number of steps or resolution _lowerCamelCase : List[Any] = [a, b] # define boundary of integration _lowerCamelCase : Optional[Any] = method_a(lowercase__ , lowercase__ ) print(f'''y = {y}''' ) if __name__ == "__main__": main()
96
"""simple docstring""" import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = ["""image_processor""", """tokenizer"""] lowerCamelCase__ = """BlipImageProcessor""" lowerCamelCase__ = """AutoTokenizer""" def __init__( self , lowercase , lowercase , lowercase ): super().__init__(lowercase , lowercase ) # add QFormer tokenizer _lowerCamelCase : int = qformer_tokenizer def __call__( self , lowercase = None , lowercase = None , lowercase = True , lowercase = False , lowercase = None , lowercase = None , lowercase = 0 , lowercase = None , lowercase = None , lowercase = False , lowercase = False , lowercase = False , lowercase = False , lowercase = False , lowercase = True , lowercase = None , **lowercase , ): if images is None and text is None: raise ValueError('You have to specify at least images or text.' ) _lowerCamelCase : int = BatchFeature() if text is not None: _lowerCamelCase : List[str] = self.tokenizer( text=lowercase , add_special_tokens=lowercase , padding=lowercase , truncation=lowercase , max_length=lowercase , stride=lowercase , pad_to_multiple_of=lowercase , return_attention_mask=lowercase , return_overflowing_tokens=lowercase , return_special_tokens_mask=lowercase , return_offsets_mapping=lowercase , return_token_type_ids=lowercase , return_length=lowercase , verbose=lowercase , return_tensors=lowercase , **lowercase , ) encoding.update(lowercase ) _lowerCamelCase : List[str] = self.qformer_tokenizer( text=lowercase , add_special_tokens=lowercase , padding=lowercase , truncation=lowercase , max_length=lowercase , stride=lowercase , pad_to_multiple_of=lowercase , return_attention_mask=lowercase , return_overflowing_tokens=lowercase , return_special_tokens_mask=lowercase , return_offsets_mapping=lowercase , return_token_type_ids=lowercase , return_length=lowercase , verbose=lowercase , return_tensors=lowercase , **lowercase , ) _lowerCamelCase : List[Any] = qformer_text_encoding.pop('input_ids' ) _lowerCamelCase : Tuple = qformer_text_encoding.pop('attention_mask' ) if images is not None: _lowerCamelCase : int = self.image_processor(lowercase , return_tensors=lowercase ) encoding.update(lowercase ) return encoding def A_ ( self , *lowercase , **lowercase ): return self.tokenizer.batch_decode(*lowercase , **lowercase ) def A_ ( self , *lowercase , **lowercase ): return self.tokenizer.decode(*lowercase , **lowercase ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def A_ ( self ): _lowerCamelCase : Union[str, Any] = self.tokenizer.model_input_names _lowerCamelCase : Any = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def A_ ( self , lowercase , **lowercase ): if os.path.isfile(lowercase ): raise ValueError(F'''Provided path ({save_directory}) should be a directory, not a file''' ) os.makedirs(lowercase , exist_ok=lowercase ) _lowerCamelCase : Optional[Any] = os.path.join(lowercase , 'qformer_tokenizer' ) self.qformer_tokenizer.save_pretrained(lowercase ) return super().save_pretrained(lowercase , **lowercase ) @classmethod def A_ ( cls , lowercase , **lowercase ): _lowerCamelCase : Union[str, Any] = AutoTokenizer.from_pretrained(lowercase , subfolder='qformer_tokenizer' ) _lowerCamelCase : Dict = cls._get_arguments_from_pretrained(lowercase , **lowercase ) args.append(lowercase ) return cls(*lowercase )
96
1
"""simple docstring""" def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): if height >= 1: move_tower(height - 1 , lowercase__ , lowercase__ , lowercase__ ) move_disk(lowercase__ , lowercase__ ) move_tower(height - 1 , lowercase__ , lowercase__ , lowercase__ ) def _snake_case ( lowercase__ , lowercase__ ): print('moving disk from' , lowercase__ , 'to' , lowercase__ ) def _snake_case ( ): _lowerCamelCase : List[Any] = int(input('Height of hanoi: ' ).strip() ) move_tower(lowercase__ , 'A' , 'B' , 'C' ) if __name__ == "__main__": main()
96
"""simple docstring""" import json import logging import os import sys from time import time from unittest.mock import patch from transformers.testing_utils import TestCasePlus, require_torch_tpu logging.basicConfig(level=logging.DEBUG) lowercase__ = logging.getLogger() def _snake_case ( lowercase__ ): _lowerCamelCase : List[Any] = {} _lowerCamelCase : List[Any] = os.path.join(lowercase__ , 'all_results.json' ) if os.path.exists(lowercase__ ): with open(lowercase__ , 'r' ) as f: _lowerCamelCase : List[Any] = json.load(lowercase__ ) else: raise ValueError(f'''can\'t find {path}''' ) return results lowercase__ = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) @require_torch_tpu class lowerCAmelCase__ ( lowercase ): '''simple docstring''' def A_ ( self ): import xla_spawn _lowerCamelCase : List[Any] = self.get_auto_remove_tmp_dir() _lowerCamelCase : List[Any] = F''' ./examples/pytorch/text-classification/run_glue.py --num_cores=8 ./examples/pytorch/text-classification/run_glue.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --overwrite_output_dir --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --do_train --do_eval --debug tpu_metrics_debug --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --max_steps=10 --warmup_steps=2 --seed=42 --max_seq_length=128 '''.split() with patch.object(lowercase , 'argv' , lowercase ): _lowerCamelCase : Dict = time() xla_spawn.main() _lowerCamelCase : Any = time() _lowerCamelCase : Optional[int] = get_results(lowercase ) self.assertGreaterEqual(result['eval_accuracy'] , 0.75 ) # Assert that the script takes less than 500 seconds to make sure it doesn't hang. self.assertLess(end - start , 500 ) def A_ ( self ): import xla_spawn _lowerCamelCase : Tuple = '\n ./tests/test_trainer_tpu.py\n --num_cores=8\n ./tests/test_trainer_tpu.py\n '.split() with patch.object(lowercase , 'argv' , lowercase ): xla_spawn.main()
96
1
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class lowerCAmelCase__ : '''simple docstring''' def __init__( self , lowercase , lowercase=13 , lowercase=2 , lowercase=24 , lowercase=16 , lowercase=True , lowercase=True , lowercase=32 , lowercase=5 , lowercase=4 , lowercase=37 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=10 , lowercase=0.02 , lowercase=None , lowercase=2 , lowercase=2 , ): _lowerCamelCase : List[Any] = parent _lowerCamelCase : Optional[int] = batch_size _lowerCamelCase : Dict = patch_size _lowerCamelCase : List[Any] = max_length _lowerCamelCase : Optional[int] = num_mel_bins _lowerCamelCase : List[str] = is_training _lowerCamelCase : Optional[int] = use_labels _lowerCamelCase : Dict = hidden_size _lowerCamelCase : Union[str, Any] = num_hidden_layers _lowerCamelCase : List[Any] = num_attention_heads _lowerCamelCase : List[Any] = intermediate_size _lowerCamelCase : str = hidden_act _lowerCamelCase : str = hidden_dropout_prob _lowerCamelCase : List[str] = attention_probs_dropout_prob _lowerCamelCase : List[str] = type_sequence_label_size _lowerCamelCase : Optional[int] = initializer_range _lowerCamelCase : Dict = scope _lowerCamelCase : List[str] = frequency_stride _lowerCamelCase : int = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) _lowerCamelCase : List[Any] = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 _lowerCamelCase : str = (self.max_length - self.patch_size) // self.time_stride + 1 _lowerCamelCase : Optional[Any] = frequency_out_dimension * time_out_dimension _lowerCamelCase : Dict = num_patches + 2 def A_ ( self ): _lowerCamelCase : List[str] = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) _lowerCamelCase : Union[str, Any] = None if self.use_labels: _lowerCamelCase : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase : Dict = self.get_config() return config, input_values, labels def A_ ( self ): return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowercase , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def A_ ( self , lowercase , lowercase , lowercase ): _lowerCamelCase : Dict = ASTModel(config=lowercase ) model.to(lowercase ) model.eval() _lowerCamelCase : str = model(lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A_ ( self ): _lowerCamelCase : List[Any] = self.prepare_config_and_inputs() ( ( _lowerCamelCase ), ( _lowerCamelCase ), ( _lowerCamelCase ), ) : List[Any] = config_and_inputs _lowerCamelCase : int = {'input_values': input_values} return config, inputs_dict @require_torch class lowerCAmelCase__ ( lowercase, lowercase, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) lowerCamelCase__ = ( {"""audio-classification""": ASTForAudioClassification, """feature-extraction""": ASTModel} if is_torch_available() else {} ) lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False def A_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase ): if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def A_ ( self ): _lowerCamelCase : Union[str, Any] = ASTModelTester(self ) _lowerCamelCase : Optional[Any] = ConfigTester(self , config_class=lowercase , has_text_modality=lowercase , hidden_size=37 ) def A_ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='AST does not use inputs_embeds' ) def A_ ( self ): pass def A_ ( self ): _lowerCamelCase, _lowerCamelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase : int = model_class(lowercase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _lowerCamelCase : List[str] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase , nn.Linear ) ) def A_ ( self ): _lowerCamelCase, _lowerCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase : List[str] = model_class(lowercase ) _lowerCamelCase : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase : Optional[Any] = [*signature.parameters.keys()] _lowerCamelCase : List[Any] = ['input_values'] self.assertListEqual(arg_names[:1] , lowercase ) def A_ ( self ): _lowerCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase ) @slow def A_ ( self ): for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : int = ASTModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) def _snake_case ( ): _lowerCamelCase : Any = hf_hub_download( repo_id='nielsr/audio-spectogram-transformer-checkpoint' , filename='sample_audio.flac' , repo_type='dataset' ) _lowerCamelCase, _lowerCamelCase : Dict = torchaudio.load(lowercase__ ) return audio, sampling_rate @require_torch @require_torchaudio class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' @cached_property def A_ ( self ): return ( ASTFeatureExtractor.from_pretrained('MIT/ast-finetuned-audioset-10-10-0.4593' ) if is_torchaudio_available() else None ) @slow def A_ ( self ): _lowerCamelCase : Optional[Any] = self.default_feature_extractor _lowerCamelCase : str = ASTForAudioClassification.from_pretrained('MIT/ast-finetuned-audioset-10-10-0.4593' ).to(lowercase ) _lowerCamelCase : List[str] = self.default_feature_extractor _lowerCamelCase, _lowerCamelCase : List[Any] = prepare_audio() _lowerCamelCase : Union[str, Any] = audio.squeeze().numpy() _lowerCamelCase : Dict = feature_extractor(lowercase , sampling_rate=lowercase , return_tensors='pt' ).to(lowercase ) # forward pass with torch.no_grad(): _lowerCamelCase : int = model(**lowercase ) # verify the logits _lowerCamelCase : Dict = torch.Size((1, 527) ) self.assertEqual(outputs.logits.shape , lowercase ) _lowerCamelCase : Optional[Any] = torch.tensor([-0.87_60, -7.00_42, -8.66_02] ).to(lowercase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowercase , atol=1E-4 ) )
96
"""simple docstring""" from __future__ import annotations import math import numpy as np from numpy.linalg import norm def _snake_case ( lowercase__ , lowercase__ ): return math.sqrt(sum(pow(a - b , 2 ) for a, b in zip(lowercase__ , lowercase__ ) ) ) def _snake_case ( lowercase__ , lowercase__ ): if dataset.ndim != value_array.ndim: _lowerCamelCase : Tuple = ( 'Wrong input data\'s dimensions... ' f'''dataset : {dataset.ndim}, value_array : {value_array.ndim}''' ) raise ValueError(lowercase__ ) try: if dataset.shape[1] != value_array.shape[1]: _lowerCamelCase : Optional[int] = ( 'Wrong input data\'s shape... ' f'''dataset : {dataset.shape[1]}, value_array : {value_array.shape[1]}''' ) raise ValueError(lowercase__ ) except IndexError: if dataset.ndim != value_array.ndim: raise TypeError('Wrong shape' ) if dataset.dtype != value_array.dtype: _lowerCamelCase : int = ( 'Input data have different datatype... ' f'''dataset : {dataset.dtype}, value_array : {value_array.dtype}''' ) raise TypeError(lowercase__ ) _lowerCamelCase : Optional[int] = [] for value in value_array: _lowerCamelCase : Tuple = euclidean(lowercase__ , dataset[0] ) _lowerCamelCase : Union[str, Any] = dataset[0].tolist() for dataset_value in dataset[1:]: _lowerCamelCase : Optional[Any] = euclidean(lowercase__ , lowercase__ ) if dist > temp_dist: _lowerCamelCase : List[Any] = temp_dist _lowerCamelCase : List[str] = dataset_value.tolist() answer.append([vector, dist] ) return answer def _snake_case ( lowercase__ , lowercase__ ): return np.dot(lowercase__ , lowercase__ ) / (norm(lowercase__ ) * norm(lowercase__ )) if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" import unittest import numpy as np from transformers import RobertaConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.roberta.modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, ) class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def __init__( self , lowercase , lowercase=13 , lowercase=7 , lowercase=True , lowercase=True , lowercase=True , lowercase=True , lowercase=99 , lowercase=32 , lowercase=5 , lowercase=4 , lowercase=37 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=512 , lowercase=16 , lowercase=2 , lowercase=0.02 , lowercase=4 , ): _lowerCamelCase : Optional[int] = parent _lowerCamelCase : Dict = batch_size _lowerCamelCase : Tuple = seq_length _lowerCamelCase : Tuple = is_training _lowerCamelCase : int = use_attention_mask _lowerCamelCase : Tuple = use_token_type_ids _lowerCamelCase : Union[str, Any] = use_labels _lowerCamelCase : Any = vocab_size _lowerCamelCase : Tuple = hidden_size _lowerCamelCase : Any = num_hidden_layers _lowerCamelCase : Optional[int] = num_attention_heads _lowerCamelCase : Dict = intermediate_size _lowerCamelCase : Tuple = hidden_act _lowerCamelCase : Union[str, Any] = hidden_dropout_prob _lowerCamelCase : Optional[int] = attention_probs_dropout_prob _lowerCamelCase : List[str] = max_position_embeddings _lowerCamelCase : Any = type_vocab_size _lowerCamelCase : Any = type_sequence_label_size _lowerCamelCase : List[str] = initializer_range _lowerCamelCase : Optional[int] = num_choices def A_ ( self ): _lowerCamelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase : Optional[int] = None if self.use_attention_mask: _lowerCamelCase : int = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCamelCase : Optional[Any] = None if self.use_token_type_ids: _lowerCamelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCamelCase : List[Any] = RobertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowercase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def A_ ( self ): _lowerCamelCase : Tuple = self.prepare_config_and_inputs() _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : List[str] = config_and_inputs _lowerCamelCase : str = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': attention_mask} return config, inputs_dict def A_ ( self ): _lowerCamelCase : Dict = self.prepare_config_and_inputs() _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : Dict = config_and_inputs _lowerCamelCase : Dict = True _lowerCamelCase : List[str] = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) _lowerCamelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax class lowerCAmelCase__ ( lowercase, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = True lowerCamelCase__ = ( ( FlaxRobertaModel, FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, ) if is_flax_available() else () ) def A_ ( self ): _lowerCamelCase : Union[str, Any] = FlaxRobertaModelTester(self ) @slow def A_ ( self ): for model_class_name in self.all_model_classes: _lowerCamelCase : Any = model_class_name.from_pretrained('roberta-base' , from_pt=lowercase ) _lowerCamelCase : List[Any] = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowercase )
96
"""simple docstring""" import socket def _snake_case ( ): _lowerCamelCase : List[Any] = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) _lowerCamelCase : Union[str, Any] = socket.gethostname() _lowerCamelCase : List[Any] = 12312 sock.connect((host, port) ) sock.send(B'Hello server!' ) with open('Received_file' , 'wb' ) as out_file: print('File opened' ) print('Receiving data...' ) while True: _lowerCamelCase : int = sock.recv(1024 ) if not data: break out_file.write(lowercase__ ) print('Successfully received the file' ) sock.close() print('Connection closed' ) if __name__ == "__main__": main()
96
1
"""simple docstring""" import argparse import hashlib # hashlib is only used inside the Test class import struct class lowerCAmelCase__ : '''simple docstring''' def __init__( self , lowercase ): _lowerCamelCase : Tuple = data _lowerCamelCase : int = [0X67452301, 0XEFCDAB89, 0X98BADCFE, 0X10325476, 0XC3D2E1F0] @staticmethod def A_ ( lowercase , lowercase ): return ((n << b) | (n >> (32 - b))) & 0XFFFFFFFF def A_ ( self ): _lowerCamelCase : List[str] = B'\x80' + B'\x00' * (63 - (len(self.data ) + 8) % 64) _lowerCamelCase : Optional[int] = self.data + padding + struct.pack('>Q' , 8 * len(self.data ) ) return padded_data def A_ ( self ): return [ self.padded_data[i : i + 64] for i in range(0 , len(self.padded_data ) , 64 ) ] def A_ ( self , lowercase ): _lowerCamelCase : Union[str, Any] = list(struct.unpack('>16L' , lowercase ) ) + [0] * 64 for i in range(16 , 80 ): _lowerCamelCase : str = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 14] ^ w[i - 16]) , 1 ) return w def A_ ( self ): _lowerCamelCase : Any = self.padding() _lowerCamelCase : str = self.split_blocks() for block in self.blocks: _lowerCamelCase : Union[str, Any] = self.expand_block(lowercase ) _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : Any = self.h for i in range(0 , 80 ): if 0 <= i < 20: _lowerCamelCase : Union[str, Any] = (b & c) | ((~b) & d) _lowerCamelCase : Optional[Any] = 0X5A827999 elif 20 <= i < 40: _lowerCamelCase : str = b ^ c ^ d _lowerCamelCase : List[str] = 0X6ED9EBA1 elif 40 <= i < 60: _lowerCamelCase : Any = (b & c) | (b & d) | (c & d) _lowerCamelCase : Optional[Any] = 0X8F1BBCDC elif 60 <= i < 80: _lowerCamelCase : Union[str, Any] = b ^ c ^ d _lowerCamelCase : List[str] = 0XCA62C1D6 _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : Any = ( self.rotate(lowercase , 5 ) + f + e + k + expanded_block[i] & 0XFFFFFFFF, a, self.rotate(lowercase , 30 ), c, d, ) _lowerCamelCase : Any = ( self.h[0] + a & 0XFFFFFFFF, self.h[1] + b & 0XFFFFFFFF, self.h[2] + c & 0XFFFFFFFF, self.h[3] + d & 0XFFFFFFFF, self.h[4] + e & 0XFFFFFFFF, ) return ("{:08x}" * 5).format(*self.h ) def _snake_case ( ): _lowerCamelCase : List[str] = B'Test String' assert SHAaHash(lowercase__ ).final_hash() == hashlib.shaa(lowercase__ ).hexdigest() # noqa: S324 def _snake_case ( ): _lowerCamelCase : Optional[int] = argparse.ArgumentParser(description='Process some strings or files' ) parser.add_argument( '--string' , dest='input_string' , default='Hello World!! Welcome to Cryptography' , help='Hash the string' , ) parser.add_argument('--file' , dest='input_file' , help='Hash contents of a file' ) _lowerCamelCase : Dict = parser.parse_args() _lowerCamelCase : List[str] = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file , 'rb' ) as f: _lowerCamelCase : Optional[Any] = f.read() else: _lowerCamelCase : List[str] = bytes(lowercase__ , 'utf-8' ) print(SHAaHash(lowercase__ ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
96
"""simple docstring""" from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record lowercase__ = """\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } """ lowercase__ = """\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. """ lowercase__ = """ Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for 'record': list of question-answer dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'prediction_text': the predicted answer text - for 'multirc': list of question-answer dictionaries with the following keys: - 'idx': index of the question-answer pair as specified by the dataset - 'prediction': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for 'record': list of question-answers dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'answers': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for 'record': - 'exact_match': Exact match between answer and gold answer - 'f1': F1 score - for 'multirc': - 'exact_match': Exact match between answer and gold answer - 'f1_m': Per-question macro-F1 score - 'f1_a': Average F1 score over all answers - for 'axb': 'matthews_correlation': Matthew Correlation - for 'cb': - 'accuracy': Accuracy - 'f1': F1 score - for all others: - 'accuracy': Accuracy Examples: >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'cb') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'record') >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}] >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc') >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'axb') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def _snake_case ( lowercase__ , lowercase__ ): return float((preds == labels).mean() ) def _snake_case ( lowercase__ , lowercase__ , lowercase__="binary" ): _lowerCamelCase : str = simple_accuracy(lowercase__ , lowercase__ ) _lowerCamelCase : Any = float(fa_score(y_true=lowercase__ , y_pred=lowercase__ , average=lowercase__ ) ) return { "accuracy": acc, "f1": fa, } def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : Any = {} for id_pred, label in zip(lowercase__ , lowercase__ ): _lowerCamelCase : Tuple = f'''{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}''' _lowerCamelCase : Union[str, Any] = id_pred['prediction'] if question_id in question_map: question_map[question_id].append((pred, label) ) else: _lowerCamelCase : Optional[Any] = [(pred, label)] _lowerCamelCase, _lowerCamelCase : Optional[int] = [], [] for question, preds_labels in question_map.items(): _lowerCamelCase, _lowerCamelCase : Tuple = zip(*lowercase__ ) _lowerCamelCase : List[str] = fa_score(y_true=lowercase__ , y_pred=lowercase__ , average='macro' ) fas.append(lowercase__ ) _lowerCamelCase : int = int(sum(pred == label for pred, label in preds_labels ) == len(lowercase__ ) ) ems.append(lowercase__ ) _lowerCamelCase : Optional[Any] = float(sum(lowercase__ ) / len(lowercase__ ) ) _lowerCamelCase : Optional[int] = sum(lowercase__ ) / len(lowercase__ ) _lowerCamelCase : List[Any] = float(fa_score(y_true=lowercase__ , y_pred=[id_pred['prediction'] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCAmelCase__ ( datasets.Metric ): '''simple docstring''' def A_ ( self ): if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( 'You should supply a configuration name selected in ' '["boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg",]' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format='numpy' if not self.config_name == 'record' and not self.config_name == 'multirc' else None , ) def A_ ( self ): if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value('int64' ), "query": datasets.Value('int64' ), }, "prediction_text": datasets.Value('string' ), }, "references": { "idx": { "passage": datasets.Value('int64' ), "query": datasets.Value('int64' ), }, "answers": datasets.Sequence(datasets.Value('string' ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value('int64' ), "paragraph": datasets.Value('int64' ), "question": datasets.Value('int64' ), }, "prediction": datasets.Value('int64' ), }, "references": datasets.Value('int64' ), } else: return { "predictions": datasets.Value('int64' ), "references": datasets.Value('int64' ), } def A_ ( self , lowercase , lowercase ): if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(lowercase , lowercase )} elif self.config_name == "cb": return acc_and_fa(lowercase , lowercase , fa_avg='macro' ) elif self.config_name == "record": _lowerCamelCase : List[str] = [ { 'qas': [ {'id': ref['idx']['query'], 'answers': [{'text': ans} for ans in ref['answers']]} for ref in references ] } ] _lowerCamelCase : Union[str, Any] = {pred['idx']['query']: pred['prediction_text'] for pred in predictions} return evaluate_record(lowercase , lowercase )[0] elif self.config_name == "multirc": return evaluate_multirc(lowercase , lowercase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(lowercase , lowercase )} else: raise KeyError( 'You should supply a configuration name selected in ' '["boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg",]' )
96
1
"""simple docstring""" from __future__ import annotations def _snake_case ( lowercase__ , lowercase__ = None , lowercase__ = None ): if start is None: _lowerCamelCase : str = 0 if end is None: _lowerCamelCase : Tuple = len(lowercase__ ) - 1 if start >= end: return _lowerCamelCase : Optional[int] = (start + end) // 2 slowsort(lowercase__ , lowercase__ , lowercase__ ) slowsort(lowercase__ , mid + 1 , lowercase__ ) if sequence[end] < sequence[mid]: _lowerCamelCase, _lowerCamelCase : Union[str, Any] = sequence[mid], sequence[end] slowsort(lowercase__ , lowercase__ , end - 1 ) if __name__ == "__main__": from doctest import testmod testmod()
96
"""simple docstring""" import unittest import numpy as np import torch from diffusers import DDIMPipeline, DDIMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow, torch_device from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class lowerCAmelCase__ ( lowercase, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = DDIMPipeline lowerCamelCase__ = UNCONDITIONAL_IMAGE_GENERATION_PARAMS lowerCamelCase__ = PipelineTesterMixin.required_optional_params - { """num_images_per_prompt""", """latents""", """callback""", """callback_steps""", } lowerCamelCase__ = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS lowerCamelCase__ = False def A_ ( self ): torch.manual_seed(0 ) _lowerCamelCase : List[Any] = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('DownBlock2D', 'AttnDownBlock2D') , up_block_types=('AttnUpBlock2D', 'UpBlock2D') , ) _lowerCamelCase : List[str] = DDIMScheduler() _lowerCamelCase : Optional[int] = {'unet': unet, 'scheduler': scheduler} return components def A_ ( self , lowercase , lowercase=0 ): if str(lowercase ).startswith('mps' ): _lowerCamelCase : Dict = torch.manual_seed(lowercase ) else: _lowerCamelCase : List[str] = torch.Generator(device=lowercase ).manual_seed(lowercase ) _lowerCamelCase : Tuple = { 'batch_size': 1, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs def A_ ( self ): _lowerCamelCase : Any = 'cpu' _lowerCamelCase : Tuple = self.get_dummy_components() _lowerCamelCase : Optional[Any] = self.pipeline_class(**lowercase ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : str = self.get_dummy_inputs(lowercase ) _lowerCamelCase : int = pipe(**lowercase ).images _lowerCamelCase : Any = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 32, 32, 3) ) _lowerCamelCase : Tuple = np.array( [1.000E00, 5.717E-01, 4.717E-01, 1.000E00, 0.000E00, 1.000E00, 3.000E-04, 0.000E00, 9.000E-04] ) _lowerCamelCase : str = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowercase , 1E-3 ) def A_ ( self ): super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) def A_ ( self ): super().test_save_load_local(expected_max_difference=3E-3 ) def A_ ( self ): super().test_save_load_optional_components(expected_max_difference=3E-3 ) def A_ ( self ): super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def A_ ( self ): _lowerCamelCase : Optional[Any] = 'google/ddpm-cifar10-32' _lowerCamelCase : Optional[Any] = UNetaDModel.from_pretrained(lowercase ) _lowerCamelCase : Dict = DDIMScheduler() _lowerCamelCase : Dict = DDIMPipeline(unet=lowercase , scheduler=lowercase ) ddim.to(lowercase ) ddim.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : List[str] = torch.manual_seed(0 ) _lowerCamelCase : str = ddim(generator=lowercase , eta=0.0 , output_type='numpy' ).images _lowerCamelCase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _lowerCamelCase : List[Any] = np.array([0.17_23, 0.16_17, 0.16_00, 0.16_26, 0.14_97, 0.15_13, 0.15_05, 0.14_42, 0.14_53] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def A_ ( self ): _lowerCamelCase : Optional[int] = 'google/ddpm-ema-bedroom-256' _lowerCamelCase : str = UNetaDModel.from_pretrained(lowercase ) _lowerCamelCase : str = DDIMScheduler.from_pretrained(lowercase ) _lowerCamelCase : Optional[int] = DDIMPipeline(unet=lowercase , scheduler=lowercase ) ddpm.to(lowercase ) ddpm.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : Tuple = torch.manual_seed(0 ) _lowerCamelCase : int = ddpm(generator=lowercase , output_type='numpy' ).images _lowerCamelCase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _lowerCamelCase : str = np.array([0.00_60, 0.02_01, 0.03_44, 0.00_24, 0.00_18, 0.00_02, 0.00_22, 0.00_00, 0.00_69] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
96
1
"""simple docstring""" import numpy as np def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : List[str] = int(np.ceil((x_end - xa) / h ) ) _lowerCamelCase : Any = np.zeros((n + 1,) ) _lowerCamelCase : Optional[int] = ya _lowerCamelCase : Dict = xa for k in range(lowercase__ ): _lowerCamelCase : Any = f(lowercase__ , y[k] ) _lowerCamelCase : str = f(x + 0.5 * h , y[k] + 0.5 * h * ka ) _lowerCamelCase : Tuple = f(x + 0.5 * h , y[k] + 0.5 * h * ka ) _lowerCamelCase : Optional[Any] = f(x + h , y[k] + h * ka ) _lowerCamelCase : Tuple = y[k] + (1 / 6) * h * (ka + 2 * ka + 2 * ka + ka) x += h return y if __name__ == "__main__": import doctest doctest.testmod()
96
"""simple docstring""" # Imports import numpy as np class lowerCAmelCase__ : '''simple docstring''' def __init__( self , lowercase=None , lowercase=None , lowercase=None , lowercase=None , lowercase=None ): self.set_matricies(red=lowercase , green=lowercase , blue=lowercase , red_edge=lowercase , nir=lowercase ) def A_ ( self , lowercase=None , lowercase=None , lowercase=None , lowercase=None , lowercase=None ): if red is not None: _lowerCamelCase : Optional[int] = red if green is not None: _lowerCamelCase : Optional[Any] = green if blue is not None: _lowerCamelCase : Tuple = blue if red_edge is not None: _lowerCamelCase : Optional[Any] = red_edge if nir is not None: _lowerCamelCase : Union[str, Any] = nir return True def A_ ( self , lowercase="" , lowercase=None , lowercase=None , lowercase=None , lowercase=None , lowercase=None ): self.set_matricies(red=lowercase , green=lowercase , blue=lowercase , red_edge=lowercase , nir=lowercase ) _lowerCamelCase : str = { 'ARVI2': self.arvaa, 'CCCI': self.ccci, 'CVI': self.cvi, 'GLI': self.gli, 'NDVI': self.ndvi, 'BNDVI': self.bndvi, 'redEdgeNDVI': self.red_edge_ndvi, 'GNDVI': self.gndvi, 'GBNDVI': self.gbndvi, 'GRNDVI': self.grndvi, 'RBNDVI': self.rbndvi, 'PNDVI': self.pndvi, 'ATSAVI': self.atsavi, 'BWDRVI': self.bwdrvi, 'CIgreen': self.ci_green, 'CIrededge': self.ci_rededge, 'CI': self.ci, 'CTVI': self.ctvi, 'GDVI': self.gdvi, 'EVI': self.evi, 'GEMI': self.gemi, 'GOSAVI': self.gosavi, 'GSAVI': self.gsavi, 'Hue': self.hue, 'IVI': self.ivi, 'IPVI': self.ipvi, 'I': self.i, 'RVI': self.rvi, 'MRVI': self.mrvi, 'MSAVI': self.m_savi, 'NormG': self.norm_g, 'NormNIR': self.norm_nir, 'NormR': self.norm_r, 'NGRDI': self.ngrdi, 'RI': self.ri, 'S': self.s, 'IF': self._if, 'DVI': self.dvi, 'TVI': self.tvi, 'NDRE': self.ndre, } try: return funcs[index]() except KeyError: print('Index not in the list!' ) return False def A_ ( self ): return -0.18 + (1.17 * ((self.nir - self.red) / (self.nir + self.red))) def A_ ( self ): return ((self.nir - self.redEdge) / (self.nir + self.redEdge)) / ( (self.nir - self.red) / (self.nir + self.red) ) def A_ ( self ): return self.nir * (self.red / (self.green**2)) def A_ ( self ): return (2 * self.green - self.red - self.blue) / ( 2 * self.green + self.red + self.blue ) def A_ ( self ): return (self.nir - self.red) / (self.nir + self.red) def A_ ( self ): return (self.nir - self.blue) / (self.nir + self.blue) def A_ ( self ): return (self.redEdge - self.red) / (self.redEdge + self.red) def A_ ( self ): return (self.nir - self.green) / (self.nir + self.green) def A_ ( self ): return (self.nir - (self.green + self.blue)) / ( self.nir + (self.green + self.blue) ) def A_ ( self ): return (self.nir - (self.green + self.red)) / ( self.nir + (self.green + self.red) ) def A_ ( self ): return (self.nir - (self.blue + self.red)) / (self.nir + (self.blue + self.red)) def A_ ( self ): return (self.nir - (self.green + self.red + self.blue)) / ( self.nir + (self.green + self.red + self.blue) ) def A_ ( self , lowercase=0.08 , lowercase=1.22 , lowercase=0.03 ): return a * ( (self.nir - a * self.red - b) / (a * self.nir + self.red - a * b + x * (1 + a**2)) ) def A_ ( self ): return (0.1 * self.nir - self.blue) / (0.1 * self.nir + self.blue) def A_ ( self ): return (self.nir / self.green) - 1 def A_ ( self ): return (self.nir / self.redEdge) - 1 def A_ ( self ): return (self.red - self.blue) / self.red def A_ ( self ): _lowerCamelCase : Any = self.ndvi() return ((ndvi + 0.5) / (abs(ndvi + 0.5 ))) * (abs(ndvi + 0.5 ) ** (1 / 2)) def A_ ( self ): return self.nir - self.green def A_ ( self ): return 2.5 * ( (self.nir - self.red) / (self.nir + 6 * self.red - 7.5 * self.blue + 1) ) def A_ ( self ): _lowerCamelCase : Any = (2 * (self.nir**2 - self.red**2) + 1.5 * self.nir + 0.5 * self.red) / ( self.nir + self.red + 0.5 ) return n * (1 - 0.25 * n) - (self.red - 0.1_25) / (1 - self.red) def A_ ( self , lowercase=0.16 ): return (self.nir - self.green) / (self.nir + self.green + y) def A_ ( self , lowercase=0.5 ): return ((self.nir - self.green) / (self.nir + self.green + n)) * (1 + n) def A_ ( self ): return np.arctan( ((2 * self.red - self.green - self.blue) / 30.5) * (self.green - self.blue) ) def A_ ( self , lowercase=None , lowercase=None ): return (self.nir - b) / (a * self.red) def A_ ( self ): return (self.nir / ((self.nir + self.red) / 2)) * (self.ndvi() + 1) def A_ ( self ): return (self.red + self.green + self.blue) / 30.5 def A_ ( self ): return self.nir / self.red def A_ ( self ): return (self.rvi() - 1) / (self.rvi() + 1) def A_ ( self ): return ( (2 * self.nir + 1) - ((2 * self.nir + 1) ** 2 - 8 * (self.nir - self.red)) ** (1 / 2) ) / 2 def A_ ( self ): return self.green / (self.nir + self.red + self.green) def A_ ( self ): return self.nir / (self.nir + self.red + self.green) def A_ ( self ): return self.red / (self.nir + self.red + self.green) def A_ ( self ): return (self.green - self.red) / (self.green + self.red) def A_ ( self ): return (self.red - self.green) / (self.red + self.green) def A_ ( self ): _lowerCamelCase : Union[str, Any] = np.max([np.max(self.red ), np.max(self.green ), np.max(self.blue )] ) _lowerCamelCase : Dict = np.min([np.min(self.red ), np.min(self.green ), np.min(self.blue )] ) return (max_value - min_value) / max_value def A_ ( self ): return (2 * self.red - self.green - self.blue) / (self.green - self.blue) def A_ ( self ): return self.nir / self.red def A_ ( self ): return (self.ndvi() + 0.5) ** (1 / 2) def A_ ( self ): return (self.nir - self.redEdge) / (self.nir + self.redEdge)
96
1
"""simple docstring""" from __future__ import annotations import math from collections.abc import Callable def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ = 100 , ): _lowerCamelCase : Any = x_start _lowerCamelCase : Optional[int] = fnc(lowercase__ ) _lowerCamelCase : str = 0.0 for _ in range(lowercase__ ): # Approximates curve as a sequence of linear lines and sums their length _lowerCamelCase : str = (x_end - x_start) / steps + xa _lowerCamelCase : Any = fnc(lowercase__ ) length += math.hypot(xa - xa , fxa - fxa ) # Increment step _lowerCamelCase : List[Any] = xa _lowerCamelCase : Dict = fxa return length if __name__ == "__main__": def _snake_case ( lowercase__ ): return math.sin(10 * x ) print("""f(x) = sin(10 * x)""") print("""The length of the curve from x = -10 to x = 10 is:""") lowercase__ = 10 while i <= 10_0000: print(F"With {i} steps: {line_length(f, -10, 10, i)}") i *= 10
96
"""simple docstring""" import torch from torch import nn from transformers import CLIPPreTrainedModel, CLIPVisionModel from ...models.attention import BasicTransformerBlock from ...utils import logging lowercase__ = logging.get_logger(__name__) # pylint: disable=invalid-name class lowerCAmelCase__ ( lowercase ): '''simple docstring''' def __init__( self , lowercase , lowercase=768 ): super().__init__(lowercase ) _lowerCamelCase : Any = proj_size _lowerCamelCase : Dict = CLIPVisionModel(lowercase ) _lowerCamelCase : List[str] = PaintByExampleMapper(lowercase ) _lowerCamelCase : Optional[Any] = nn.LayerNorm(config.hidden_size ) _lowerCamelCase : int = nn.Linear(config.hidden_size , self.proj_size ) # uncondition for scaling _lowerCamelCase : str = nn.Parameter(torch.randn((1, 1, self.proj_size) ) ) def A_ ( self , lowercase , lowercase=False ): _lowerCamelCase : Union[str, Any] = self.model(pixel_values=lowercase ) _lowerCamelCase : int = clip_output.pooler_output _lowerCamelCase : str = self.mapper(latent_states[:, None] ) _lowerCamelCase : List[Any] = self.final_layer_norm(lowercase ) _lowerCamelCase : Dict = self.proj_out(lowercase ) if return_uncond_vector: return latent_states, self.uncond_vector return latent_states class lowerCAmelCase__ ( nn.Module ): '''simple docstring''' def __init__( self , lowercase ): super().__init__() _lowerCamelCase : Tuple = (config.num_hidden_layers + 1) // 5 _lowerCamelCase : int = config.hidden_size _lowerCamelCase : Optional[Any] = 1 _lowerCamelCase : str = nn.ModuleList( [ BasicTransformerBlock(lowercase , lowercase , lowercase , activation_fn='gelu' , attention_bias=lowercase ) for _ in range(lowercase ) ] ) def A_ ( self , lowercase ): for block in self.blocks: _lowerCamelCase : Tuple = block(lowercase ) return hidden_states
96
1
"""simple docstring""" import fire from utils import calculate_rouge, save_json def _snake_case ( lowercase__ , lowercase__ , lowercase__=None , **lowercase__ ): _lowerCamelCase : Dict = [x.strip() for x in open(lowercase__ ).readlines()] _lowerCamelCase : int = [x.strip() for x in open(lowercase__ ).readlines()][: len(lowercase__ )] _lowerCamelCase : int = calculate_rouge(lowercase__ , lowercase__ , **lowercase__ ) if save_path is not None: save_json(lowercase__ , lowercase__ , indent=lowercase__ ) return metrics # these print nicely if __name__ == "__main__": fire.Fire(calculate_rouge_path)
96
"""simple docstring""" lowercase__ = { """meter""": """m""", """kilometer""": """km""", """megametre""": """Mm""", """gigametre""": """Gm""", """terametre""": """Tm""", """petametre""": """Pm""", """exametre""": """Em""", """zettametre""": """Zm""", """yottametre""": """Ym""", } # Exponent of the factor(meter) lowercase__ = { """m""": 0, """km""": 3, """Mm""": 6, """Gm""": 9, """Tm""": 12, """Pm""": 15, """Em""": 18, """Zm""": 21, """Ym""": 24, } def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : List[Any] = from_type.lower().strip('s' ) _lowerCamelCase : List[Any] = to_type.lower().strip('s' ) _lowerCamelCase : Optional[int] = UNIT_SYMBOL.get(lowercase__ , lowercase__ ) _lowerCamelCase : Any = UNIT_SYMBOL.get(lowercase__ , lowercase__ ) if from_sanitized not in METRIC_CONVERSION: _lowerCamelCase : Tuple = ( f'''Invalid \'from_type\' value: {from_type!r}.\n''' f'''Conversion abbreviations are: {', '.join(lowercase__ )}''' ) raise ValueError(lowercase__ ) if to_sanitized not in METRIC_CONVERSION: _lowerCamelCase : Any = ( f'''Invalid \'to_type\' value: {to_type!r}.\n''' f'''Conversion abbreviations are: {', '.join(lowercase__ )}''' ) raise ValueError(lowercase__ ) _lowerCamelCase : List[Any] = METRIC_CONVERSION[from_sanitized] _lowerCamelCase : int = METRIC_CONVERSION[to_sanitized] _lowerCamelCase : List[str] = 1 if from_exponent > to_exponent: _lowerCamelCase : List[str] = from_exponent - to_exponent else: _lowerCamelCase : List[Any] = -(to_exponent - from_exponent) return value * pow(10 , lowercase__ ) if __name__ == "__main__": from doctest import testmod testmod()
96
1
"""simple docstring""" import copy import inspect import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import TimesformerConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, TimesformerForVideoClassification, TimesformerModel, ) from transformers.models.timesformer.modeling_timesformer import TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from transformers import VideoMAEImageProcessor class lowerCAmelCase__ : '''simple docstring''' def __init__( self , lowercase , lowercase=13 , lowercase=10 , lowercase=3 , lowercase=2 , lowercase=2 , lowercase=True , lowercase=True , lowercase=32 , lowercase=5 , lowercase=4 , lowercase=37 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=10 , lowercase=0.02 , lowercase="divided_space_time" , lowercase=None , ): _lowerCamelCase : Tuple = parent _lowerCamelCase : Any = batch_size _lowerCamelCase : Tuple = image_size _lowerCamelCase : str = num_channels _lowerCamelCase : Optional[Any] = patch_size _lowerCamelCase : Any = num_frames _lowerCamelCase : List[Any] = is_training _lowerCamelCase : List[Any] = use_labels _lowerCamelCase : List[Any] = hidden_size _lowerCamelCase : Optional[Any] = num_hidden_layers _lowerCamelCase : List[Any] = num_attention_heads _lowerCamelCase : List[str] = intermediate_size _lowerCamelCase : Any = hidden_act _lowerCamelCase : int = hidden_dropout_prob _lowerCamelCase : Any = attention_probs_dropout_prob _lowerCamelCase : List[str] = attention_type _lowerCamelCase : int = initializer_range _lowerCamelCase : Any = scope _lowerCamelCase : int = num_labels # in TimeSformer, the number of spatial tokens equals num_frames * num_patches per frame + 1 CLS token _lowerCamelCase : Optional[Any] = (image_size // patch_size) ** 2 _lowerCamelCase : Dict = (num_frames) * self.num_patches_per_frame + 1 def A_ ( self ): _lowerCamelCase : str = floats_tensor( [self.batch_size, self.num_frames, self.num_channels, self.image_size, self.image_size] ) _lowerCamelCase : Union[str, Any] = None if self.use_labels: _lowerCamelCase : Dict = ids_tensor([self.batch_size] , self.num_labels ) _lowerCamelCase : int = self.get_config() return config, pixel_values, labels def A_ ( self ): _lowerCamelCase : str = TimesformerConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_frames=self.num_frames , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , attention_type=self.attention_type , ) _lowerCamelCase : List[str] = self.num_labels return config def A_ ( self , lowercase , lowercase , lowercase ): _lowerCamelCase : Union[str, Any] = TimesformerModel(config=lowercase ) model.to(lowercase ) model.eval() _lowerCamelCase : Any = model(lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A_ ( self , lowercase , lowercase , lowercase ): _lowerCamelCase : Optional[Any] = TimesformerForVideoClassification(lowercase ) model.to(lowercase ) model.eval() _lowerCamelCase : List[Any] = model(lowercase ) # verify the logits shape _lowerCamelCase : List[Any] = torch.Size((self.batch_size, self.num_labels) ) self.parent.assertEqual(result.logits.shape , lowercase ) def A_ ( self ): _lowerCamelCase : Optional[int] = self.prepare_config_and_inputs() _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : List[str] = config_and_inputs _lowerCamelCase : Dict = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class lowerCAmelCase__ ( lowercase, lowercase, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = (TimesformerModel, TimesformerForVideoClassification) if is_torch_available() else () lowerCamelCase__ = ( {"""feature-extraction""": TimesformerModel, """video-classification""": TimesformerForVideoClassification} if is_torch_available() else {} ) lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False def A_ ( self ): _lowerCamelCase : Union[str, Any] = TimesformerModelTester(self ) _lowerCamelCase : List[Any] = ConfigTester( self , config_class=lowercase , has_text_modality=lowercase , hidden_size=37 ) def A_ ( self , lowercase , lowercase , lowercase=False ): _lowerCamelCase : Tuple = copy.deepcopy(lowercase ) if return_labels: if model_class in get_values(lowercase ): _lowerCamelCase : Optional[int] = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowercase ) return inputs_dict def A_ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='TimeSformer does not use inputs_embeds' ) def A_ ( self ): pass def A_ ( self ): _lowerCamelCase, _lowerCamelCase : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase : Union[str, Any] = model_class(lowercase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _lowerCamelCase : List[str] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase , nn.Linear ) ) def A_ ( self ): _lowerCamelCase, _lowerCamelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase : Optional[int] = model_class(lowercase ) _lowerCamelCase : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase : Dict = [*signature.parameters.keys()] _lowerCamelCase : Optional[int] = ['pixel_values'] self.assertListEqual(arg_names[:1] , lowercase ) def A_ ( self ): _lowerCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase ) def A_ ( self ): _lowerCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_video_classification(*lowercase ) @slow def A_ ( self ): for model_name in TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : str = TimesformerModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) def A_ ( self ): if not self.has_attentions: pass else: _lowerCamelCase, _lowerCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase : Union[str, Any] = True for model_class in self.all_model_classes: _lowerCamelCase : List[Any] = self.model_tester.seq_length _lowerCamelCase : List[str] = self.model_tester.num_frames _lowerCamelCase : Dict = True _lowerCamelCase : List[Any] = False _lowerCamelCase : Optional[Any] = True _lowerCamelCase : Optional[int] = model_class(lowercase ) model.to(lowercase ) model.eval() with torch.no_grad(): _lowerCamelCase : List[Any] = model(**self._prepare_for_class(lowercase , lowercase ) ) _lowerCamelCase : Optional[int] = outputs.attentions self.assertEqual(len(lowercase ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] _lowerCamelCase : int = True _lowerCamelCase : Optional[int] = model_class(lowercase ) model.to(lowercase ) model.eval() with torch.no_grad(): _lowerCamelCase : List[Any] = model(**self._prepare_for_class(lowercase , lowercase ) ) _lowerCamelCase : List[Any] = outputs.attentions self.assertEqual(len(lowercase ) , self.model_tester.num_hidden_layers ) # attentions has shape (batch_size x num_frames) x num_heads x (num_patches per frame + 1) x (num_patches per frame + 1) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len // num_frames + 1, seq_len // num_frames + 1] , ) _lowerCamelCase : Any = len(lowercase ) # Check attention is always last and order is fine _lowerCamelCase : Optional[Any] = True _lowerCamelCase : Union[str, Any] = True _lowerCamelCase : Union[str, Any] = model_class(lowercase ) model.to(lowercase ) model.eval() with torch.no_grad(): _lowerCamelCase : str = model(**self._prepare_for_class(lowercase , lowercase ) ) self.assertEqual(out_len + 1 , len(lowercase ) ) _lowerCamelCase : Optional[Any] = outputs.attentions self.assertEqual(len(lowercase ) , self.model_tester.num_hidden_layers ) # attentions has shape (batch_size x num_frames) x num_heads x (num_patches per frame + 1) x (num_patches per frame + 1) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len // num_frames + 1, seq_len // num_frames + 1] , ) def A_ ( self ): def check_hidden_states_output(lowercase , lowercase , lowercase ): _lowerCamelCase : Any = model_class(lowercase ) model.to(lowercase ) model.eval() with torch.no_grad(): _lowerCamelCase : List[Any] = model(**self._prepare_for_class(lowercase , lowercase ) ) _lowerCamelCase : int = outputs.hidden_states _lowerCamelCase : str = self.model_tester.num_hidden_layers + 1 self.assertEqual(len(lowercase ) , lowercase ) _lowerCamelCase : Optional[Any] = self.model_tester.seq_length self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) _lowerCamelCase, _lowerCamelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase : Tuple = True check_hidden_states_output(lowercase , lowercase , lowercase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowerCamelCase : Dict = True check_hidden_states_output(lowercase , lowercase , lowercase ) def _snake_case ( ): _lowerCamelCase : str = hf_hub_download( repo_id='hf-internal-testing/spaghetti-video' , filename='eating_spaghetti.npy' , repo_type='dataset' ) _lowerCamelCase : Optional[Any] = np.load(lowercase__ ) return list(lowercase__ ) @require_torch @require_vision class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' @cached_property def A_ ( self ): # logits were tested with a different mean and std, so we use the same here return ( VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5] , image_std=[0.5, 0.5, 0.5] ) if is_vision_available() else None ) @slow def A_ ( self ): _lowerCamelCase : List[Any] = TimesformerForVideoClassification.from_pretrained('facebook/timesformer-base-finetuned-k400' ).to( lowercase ) _lowerCamelCase : Any = self.default_image_processor _lowerCamelCase : Union[str, Any] = prepare_video() _lowerCamelCase : Any = image_processor(video[:8] , return_tensors='pt' ).to(lowercase ) # forward pass with torch.no_grad(): _lowerCamelCase : Any = model(**lowercase ) # verify the logits _lowerCamelCase : Dict = torch.Size((1, 400) ) self.assertEqual(outputs.logits.shape , lowercase ) _lowerCamelCase : Optional[Any] = torch.tensor([-0.30_16, -0.77_13, -0.42_05] ).to(lowercase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowercase , atol=1E-4 ) )
96
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ = {"""configuration_vit_msn""": ["""VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ViTMSNConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ """VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST""", """ViTMSNModel""", """ViTMSNForImageClassification""", """ViTMSNPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_vit_msn import VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMSNConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_msn import ( VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMSNForImageClassification, ViTMSNModel, ViTMSNPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
96
1
"""simple docstring""" def _snake_case ( lowercase__ ): _lowerCamelCase : Optional[Any] = 1 _lowerCamelCase : Tuple = 2 while i * i <= n: _lowerCamelCase : Optional[int] = 0 while n % i == 0: n //= i multiplicity += 1 n_divisors *= multiplicity + 1 i += 1 if n > 1: n_divisors *= 2 return n_divisors def _snake_case ( ): _lowerCamelCase : str = 1 _lowerCamelCase : Any = 1 while True: i += 1 t_num += i if count_divisors(lowercase__ ) > 500: break return t_num if __name__ == "__main__": print(solution())
96
"""simple docstring""" def _snake_case ( lowercase__ , lowercase__ ): # "extended trapezoidal rule" # int(f) = dx/2 * (f1 + 2f2 + ... + fn) _lowerCamelCase : List[Any] = (boundary[1] - boundary[0]) / steps _lowerCamelCase : Tuple = boundary[0] _lowerCamelCase : Dict = boundary[1] _lowerCamelCase : List[Any] = make_points(lowercase__ , lowercase__ , lowercase__ ) _lowerCamelCase : List[Any] = 0.0 y += (h / 2.0) * f(lowercase__ ) for i in x_i: # print(i) y += h * f(lowercase__ ) y += (h / 2.0) * f(lowercase__ ) return y def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : str = a + h while x < (b - h): yield x _lowerCamelCase : int = x + h def _snake_case ( lowercase__ ): # enter your function here _lowerCamelCase : Optional[Any] = (x - 0) * (x - 0) return y def _snake_case ( ): _lowerCamelCase : int = 0.0 # Lower bound of integration _lowerCamelCase : Optional[int] = 1.0 # Upper bound of integration _lowerCamelCase : List[str] = 1_0.0 # define number of steps or resolution _lowerCamelCase : List[Any] = [a, b] # define boundary of integration _lowerCamelCase : Optional[Any] = method_a(lowercase__ , lowercase__ ) print(f'''y = {y}''' ) if __name__ == "__main__": main()
96
1
"""simple docstring""" import math def _snake_case ( lowercase__ ): if not isinstance(lowercase__ , lowercase__ ): _lowerCamelCase : Any = f'''Input value of [number={number}] must be an integer''' raise TypeError(lowercase__ ) if number < 1: _lowerCamelCase : Tuple = f'''Input value of [number={number}] must be > 0''' raise ValueError(lowercase__ ) elif number == 1: return 3 elif number == 2: return 5 else: _lowerCamelCase : Optional[Any] = int(math.log(number // 3 , 2 ) ) + 2 _lowerCamelCase : int = [3, 5] _lowerCamelCase : Dict = 2 _lowerCamelCase : Optional[int] = 3 for block in range(1 , lowercase__ ): for _ in range(lowercase__ ): proth_list.append(2 ** (block + 1) + proth_list[proth_index - 1] ) proth_index += 1 increment *= 2 return proth_list[number - 1] if __name__ == "__main__": import doctest doctest.testmod() for number in range(11): lowercase__ = 0 try: lowercase__ = proth(number) except ValueError: print(F"ValueError: there is no {number}th Proth number") continue print(F"The {number}th Proth number: {value}")
96
"""simple docstring""" import math def _snake_case ( lowercase__ ): return math.sqrt(lowercase__ ) * math.sqrt(lowercase__ ) == num def _snake_case ( lowercase__ ): _lowerCamelCase : Optional[int] = 0 _lowerCamelCase : List[Any] = n while left <= right: _lowerCamelCase : str = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: _lowerCamelCase : str = mid - 1 else: _lowerCamelCase : Optional[int] = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = {"""ctrl""": """https://huggingface.co/ctrl/resolve/main/config.json"""} class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """ctrl""" lowerCamelCase__ = ["""past_key_values"""] lowerCamelCase__ = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowercase=246534 , lowercase=256 , lowercase=1280 , lowercase=8192 , lowercase=48 , lowercase=16 , lowercase=0.1 , lowercase=0.1 , lowercase=1E-6 , lowercase=0.02 , lowercase=True , **lowercase , ): _lowerCamelCase : Any = vocab_size _lowerCamelCase : Dict = n_positions _lowerCamelCase : Optional[int] = n_embd _lowerCamelCase : str = n_layer _lowerCamelCase : Union[str, Any] = n_head _lowerCamelCase : Any = dff _lowerCamelCase : int = resid_pdrop _lowerCamelCase : Dict = embd_pdrop _lowerCamelCase : Union[str, Any] = layer_norm_epsilon _lowerCamelCase : Tuple = initializer_range _lowerCamelCase : str = use_cache super().__init__(**lowercase )
96
"""simple docstring""" import functools from typing import Any def _snake_case ( lowercase__ , lowercase__ ): # Validation if not isinstance(lowercase__ , lowercase__ ) or len(lowercase__ ) == 0: raise ValueError('the string should be not empty string' ) if not isinstance(lowercase__ , lowercase__ ) or not all( isinstance(lowercase__ , lowercase__ ) and len(lowercase__ ) > 0 for item in words ): raise ValueError('the words should be a list of non-empty strings' ) # Build trie _lowerCamelCase : dict[str, Any] = {} _lowerCamelCase : List[Any] = 'WORD_KEEPER' for word in words: _lowerCamelCase : Dict = trie for c in word: if c not in trie_node: _lowerCamelCase : Any = {} _lowerCamelCase : str = trie_node[c] _lowerCamelCase : Optional[Any] = True _lowerCamelCase : Dict = len(lowercase__ ) # Dynamic programming method @functools.cache def is_breakable(lowercase__ ) -> bool: if index == len_string: return True _lowerCamelCase : List[Any] = trie for i in range(lowercase__ , lowercase__ ): _lowerCamelCase : Any = trie_node.get(string[i] , lowercase__ ) if trie_node is None: return False if trie_node.get(lowercase__ , lowercase__ ) and is_breakable(i + 1 ): return True return False return is_breakable(0 ) if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices lowercase__ = logging.get_logger(__name__) lowercase__ = { """microsoft/focalnet-tiny""": """https://huggingface.co/microsoft/focalnet-tiny/resolve/main/config.json""", } class lowerCAmelCase__ ( lowercase, lowercase ): '''simple docstring''' lowerCamelCase__ = """focalnet""" def __init__( self , lowercase=224 , lowercase=4 , lowercase=3 , lowercase=96 , lowercase=False , lowercase=[192, 384, 768, 768] , lowercase=[2, 2, 6, 2] , lowercase=[2, 2, 2, 2] , lowercase=[3, 3, 3, 3] , lowercase="gelu" , lowercase=4.0 , lowercase=0.0 , lowercase=0.1 , lowercase=False , lowercase=1E-4 , lowercase=False , lowercase=False , lowercase=False , lowercase=0.02 , lowercase=1E-5 , lowercase=32 , lowercase=None , lowercase=None , **lowercase , ): super().__init__(**lowercase ) _lowerCamelCase : Optional[int] = image_size _lowerCamelCase : Union[str, Any] = patch_size _lowerCamelCase : int = num_channels _lowerCamelCase : Any = embed_dim _lowerCamelCase : List[Any] = use_conv_embed _lowerCamelCase : List[str] = hidden_sizes _lowerCamelCase : int = depths _lowerCamelCase : Optional[int] = focal_levels _lowerCamelCase : Any = focal_windows _lowerCamelCase : str = hidden_act _lowerCamelCase : int = mlp_ratio _lowerCamelCase : Tuple = hidden_dropout_prob _lowerCamelCase : Any = drop_path_rate _lowerCamelCase : List[Any] = use_layerscale _lowerCamelCase : Optional[Any] = layerscale_value _lowerCamelCase : Union[str, Any] = use_post_layernorm _lowerCamelCase : Tuple = use_post_layernorm_in_modulation _lowerCamelCase : Union[str, Any] = normalize_modulator _lowerCamelCase : Union[str, Any] = initializer_range _lowerCamelCase : Any = layer_norm_eps _lowerCamelCase : int = encoder_stride _lowerCamelCase : Tuple = ['stem'] + [F'''stage{idx}''' for idx in range(1 , len(self.depths ) + 1 )] _lowerCamelCase, _lowerCamelCase : Any = get_aligned_output_features_output_indices( out_features=lowercase , out_indices=lowercase , stage_names=self.stage_names )
96
"""simple docstring""" def _snake_case ( lowercase__ ): if not isinstance(lowercase__ , lowercase__ ): raise ValueError('Input series is not valid, valid series - [2, 4, 6]' ) if len(lowercase__ ) == 0: raise ValueError('Input list must be a non empty list' ) if len(lowercase__ ) == 1: return True _lowerCamelCase : List[Any] = series[1] - series[0] for index in range(len(lowercase__ ) - 1 ): if series[index + 1] - series[index] != common_diff: return False return True def _snake_case ( lowercase__ ): if not isinstance(lowercase__ , lowercase__ ): raise ValueError('Input series is not valid, valid series - [2, 4, 6]' ) if len(lowercase__ ) == 0: raise ValueError('Input list must be a non empty list' ) _lowerCamelCase : Optional[int] = 0 for val in series: answer += val return answer / len(lowercase__ ) if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" import math def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : List[Any] = len(lowercase__ ) _lowerCamelCase : Optional[Any] = int(math.floor(math.sqrt(lowercase__ ) ) ) _lowerCamelCase : Optional[int] = 0 while arr[min(lowercase__ , lowercase__ ) - 1] < x: _lowerCamelCase : Tuple = step step += int(math.floor(math.sqrt(lowercase__ ) ) ) if prev >= n: return -1 while arr[prev] < x: _lowerCamelCase : str = prev + 1 if prev == min(lowercase__ , lowercase__ ): return -1 if arr[prev] == x: return prev return -1 if __name__ == "__main__": lowercase__ = input("""Enter numbers separated by a comma:\n""").strip() lowercase__ = [int(item) for item in user_input.split(""",""")] lowercase__ = int(input("""Enter the number to be searched:\n""")) lowercase__ = jump_search(arr, x) if res == -1: print("""Number not found!""") else: print(F"Number {x} is at index {res}")
96
"""simple docstring""" import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler lowercase__ = 16 lowercase__ = 32 def _snake_case ( lowercase__ , lowercase__ = 16 , lowercase__ = "bert-base-cased" ): _lowerCamelCase : List[Any] = AutoTokenizer.from_pretrained(lowercase__ ) _lowerCamelCase : Tuple = load_dataset('glue' , 'mrpc' ) def tokenize_function(lowercase__ ): # max_length=None => use the model max length (it's actually the default) _lowerCamelCase : Union[str, Any] = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=lowercase__ , max_length=lowercase__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset _lowerCamelCase : int = datasets.map( lowercase__ , batched=lowercase__ , remove_columns=['idx', 'sentence1', 'sentence2'] , load_from_cache_file=lowercase__ ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _lowerCamelCase : Optional[int] = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(lowercase__ ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(lowercase__ , padding='max_length' , max_length=128 , return_tensors='pt' ) return tokenizer.pad(lowercase__ , padding='longest' , return_tensors='pt' ) # Instantiate dataloaders. _lowerCamelCase : List[str] = DataLoader( tokenized_datasets['train'] , shuffle=lowercase__ , collate_fn=lowercase__ , batch_size=lowercase__ ) _lowerCamelCase : int = DataLoader( tokenized_datasets['validation'] , shuffle=lowercase__ , collate_fn=lowercase__ , batch_size=lowercase__ ) return train_dataloader, eval_dataloader def _snake_case ( lowercase__ , lowercase__ ): # Initialize accelerator _lowerCamelCase : Optional[int] = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _lowerCamelCase : Optional[int] = config['lr'] _lowerCamelCase : Optional[int] = int(config['num_epochs'] ) _lowerCamelCase : Union[str, Any] = int(config['seed'] ) _lowerCamelCase : Optional[int] = int(config['batch_size'] ) _lowerCamelCase : Dict = args.model_name_or_path set_seed(lowercase__ ) _lowerCamelCase, _lowerCamelCase : Optional[int] = get_dataloaders(lowercase__ , lowercase__ , lowercase__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _lowerCamelCase : int = AutoModelForSequenceClassification.from_pretrained(lowercase__ , return_dict=lowercase__ ) # Instantiate optimizer _lowerCamelCase : Optional[int] = ( AdamW if accelerator.state.deepspeed_plugin is None or 'optimizer' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) _lowerCamelCase : Union[str, Any] = optimizer_cls(params=model.parameters() , lr=lowercase__ ) if accelerator.state.deepspeed_plugin is not None: _lowerCamelCase : str = accelerator.state.deepspeed_plugin.deepspeed_config[ 'gradient_accumulation_steps' ] else: _lowerCamelCase : Tuple = 1 _lowerCamelCase : List[Any] = (len(lowercase__ ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): _lowerCamelCase : Tuple = get_linear_schedule_with_warmup( optimizer=lowercase__ , num_warmup_steps=0 , num_training_steps=lowercase__ , ) else: _lowerCamelCase : Any = DummyScheduler(lowercase__ , total_num_steps=lowercase__ , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : Union[str, Any] = accelerator.prepare( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) # We need to keep track of how many total steps we have iterated over _lowerCamelCase : Union[str, Any] = 0 # We also need to keep track of the stating epoch so files are named properly _lowerCamelCase : Dict = 0 # Now we train the model _lowerCamelCase : Dict = evaluate.load('glue' , 'mrpc' ) _lowerCamelCase : Optional[int] = 0 _lowerCamelCase : str = {} for epoch in range(lowercase__ , lowercase__ ): model.train() for step, batch in enumerate(lowercase__ ): _lowerCamelCase : List[Any] = model(**lowercase__ ) _lowerCamelCase : int = outputs.loss _lowerCamelCase : Dict = loss / gradient_accumulation_steps accelerator.backward(lowercase__ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 model.eval() _lowerCamelCase : Union[str, Any] = 0 for step, batch in enumerate(lowercase__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _lowerCamelCase : Optional[int] = model(**lowercase__ ) _lowerCamelCase : Dict = outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times _lowerCamelCase, _lowerCamelCase : List[str] = accelerator.gather( (predictions, batch['labels']) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(lowercase__ ) - 1: _lowerCamelCase : Optional[Any] = predictions[: len(eval_dataloader.dataset ) - samples_seen] _lowerCamelCase : Dict = references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=lowercase__ , references=lowercase__ , ) _lowerCamelCase : List[Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f'''epoch {epoch}:''' , lowercase__ ) _lowerCamelCase : Tuple = eval_metric['accuracy'] if best_performance < eval_metric["accuracy"]: _lowerCamelCase : str = eval_metric['accuracy'] if args.performance_lower_bound is not None: assert ( args.performance_lower_bound <= best_performance ), f'''Best performance metric {best_performance} is lower than the lower bound {args.performance_lower_bound}''' accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , 'all_results.json' ) , 'w' ) as f: json.dump(lowercase__ , lowercase__ ) def _snake_case ( ): _lowerCamelCase : Any = argparse.ArgumentParser(description='Simple example of training script tracking peak GPU memory usage.' ) parser.add_argument( '--model_name_or_path' , type=lowercase__ , default='bert-base-cased' , help='Path to pretrained model or model identifier from huggingface.co/models.' , required=lowercase__ , ) parser.add_argument( '--output_dir' , type=lowercase__ , default='.' , help='Optional save directory where all checkpoint folders will be stored. Default is the current working directory.' , ) parser.add_argument( '--performance_lower_bound' , type=lowercase__ , default=lowercase__ , help='Optional lower bound for the performance metric. If set, the training will throw error when the performance metric drops below this value.' , ) parser.add_argument( '--num_epochs' , type=lowercase__ , default=3 , help='Number of train epochs.' , ) _lowerCamelCase : Optional[Any] = parser.parse_args() _lowerCamelCase : str = {'lr': 2E-5, 'num_epochs': args.num_epochs, 'seed': 42, 'batch_size': 16} training_function(lowercase__ , lowercase__ ) if __name__ == "__main__": main()
96
1
"""simple docstring""" from typing import Any def _snake_case ( lowercase__ ): if not input_list: return [] _lowerCamelCase : Dict = [input_list.count(lowercase__ ) for value in input_list] _lowerCamelCase : List[str] = max(lowercase__ ) # Gets the maximum count in the input list. # Gets values of modes return sorted({input_list[i] for i, value in enumerate(lowercase__ ) if value == y} ) if __name__ == "__main__": import doctest doctest.testmod()
96
"""simple docstring""" from __future__ import annotations import copy import tempfile import unittest from transformers import CONFIG_MAPPING, AutoConfig, BertConfig, GPTaConfig, TaConfig, TapasConfig, is_tf_available from transformers.testing_utils import ( DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tensorflow_probability, require_tf, slow, ) from ..bert.test_modeling_bert import BertModelTester if is_tf_available(): from transformers import ( TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelForTableQuestionAnswering, TFAutoModelForTokenClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFFunnelBaseModel, TFFunnelModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, TFTapasForQuestionAnswering, ) from transformers.models.auto.modeling_tf_auto import ( TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_MAPPING, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.tapas.modeling_tf_tapas import TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """new-model""" if is_tf_available(): class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = NewModelConfig @require_tf class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' @slow def A_ ( self ): _lowerCamelCase : List[str] = 'bert-base-cased' _lowerCamelCase : Union[str, Any] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Union[str, Any] = TFAutoModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): _lowerCamelCase : List[str] = 'bert-base-cased' _lowerCamelCase : int = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : int = TFAutoModelForPreTraining.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Dict = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : int = TFAutoModelForCausalLM.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : str = TFAutoModelForCausalLM.from_pretrained(lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : List[Any] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : str = TFAutoModelWithLMHead.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Tuple = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Optional[int] = TFAutoModelForMaskedLM.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : Tuple = TFAutoModelForMaskedLM.from_pretrained(lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Optional[int] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Tuple = TFAutoModelForSeqaSeqLM.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : Tuple = TFAutoModelForSeqaSeqLM.from_pretrained(lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: _lowerCamelCase : str = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Union[str, Any] = TFAutoModelForSequenceClassification.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: _lowerCamelCase : Optional[Any] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : List[str] = TFAutoModelForQuestionAnswering.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow @require_tensorflow_probability def A_ ( self ): for model_name in TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST[5:6]: _lowerCamelCase : Dict = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Dict = TFAutoModelForTableQuestionAnswering.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : List[Any] = TFAutoModelForTableQuestionAnswering.from_pretrained( lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) def A_ ( self ): _lowerCamelCase : int = TFAutoModelWithLMHead.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowercase ) , 14410 ) def A_ ( self ): _lowerCamelCase : Any = TFAutoModelWithLMHead.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowercase ) , 14410 ) def A_ ( self ): # For the auto model mapping, FunnelConfig has two models: FunnelModel and FunnelBaseModel _lowerCamelCase : List[str] = TFAutoModel.from_pretrained('sgugger/funnel-random-tiny' ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Optional[int] = copy.deepcopy(model.config ) _lowerCamelCase : Dict = ['FunnelBaseModel'] _lowerCamelCase : List[Any] = TFAutoModel.from_config(lowercase ) self.assertIsInstance(lowercase , lowercase ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowercase ) _lowerCamelCase : Tuple = TFAutoModel.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) def A_ ( self ): try: AutoConfig.register('new-model' , lowercase ) _lowerCamelCase : Tuple = [ TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSequenceClassification, TFAutoModelForTokenClassification, ] for auto_class in auto_classes: with self.subTest(auto_class.__name__ ): # Wrong config class will raise an error with self.assertRaises(lowercase ): auto_class.register(lowercase , lowercase ) auto_class.register(lowercase , lowercase ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(lowercase ): auto_class.register(lowercase , lowercase ) # Now that the config is registered, it can be used as any other config with the auto-API _lowerCamelCase : Optional[Any] = BertModelTester(self ).get_config() _lowerCamelCase : Dict = NewModelConfig(**tiny_config.to_dict() ) _lowerCamelCase : int = auto_class.from_config(lowercase ) self.assertIsInstance(lowercase , lowercase ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowercase ) _lowerCamelCase : List[Any] = auto_class.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"] for mapping in ( TF_MODEL_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, ): if NewModelConfig in mapping._extra_content: del mapping._extra_content[NewModelConfig] def A_ ( self ): with self.assertRaisesRegex( lowercase , 'bert-base is not a local folder and is not a valid model identifier' ): _lowerCamelCase : Union[str, Any] = TFAutoModel.from_pretrained('bert-base' ) def A_ ( self ): with self.assertRaisesRegex( lowercase , r'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): _lowerCamelCase : str = TFAutoModel.from_pretrained(lowercase , revision='aaaaaa' ) def A_ ( self ): with self.assertRaisesRegex( lowercase , 'hf-internal-testing/config-no-model does not appear to have a file named pytorch_model.bin' , ): _lowerCamelCase : Tuple = TFAutoModel.from_pretrained('hf-internal-testing/config-no-model' ) def A_ ( self ): with self.assertRaisesRegex(lowercase , 'Use `from_pt=True` to load this model' ): _lowerCamelCase : Tuple = TFAutoModel.from_pretrained('hf-internal-testing/tiny-bert-pt-only' ) def A_ ( self ): # Make sure we have cached the model. _lowerCamelCase : Optional[int] = TFAutoModel.from_pretrained('hf-internal-testing/tiny-random-bert' ) with RequestCounter() as counter: _lowerCamelCase : Optional[int] = TFAutoModel.from_pretrained('hf-internal-testing/tiny-random-bert' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 ) # With a sharded checkpoint _lowerCamelCase : int = TFAutoModel.from_pretrained('ArthurZ/tiny-random-bert-sharded' ) with RequestCounter() as counter: _lowerCamelCase : List[Any] = TFAutoModel.from_pretrained('ArthurZ/tiny-random-bert-sharded' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 )
96
1
"""simple docstring""" import random def _snake_case ( lowercase__ , lowercase__ , lowercase__ = False ): _lowerCamelCase : dict = {i: [] for i in range(lowercase__ )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(lowercase__ ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(lowercase__ ): for j in range(i + 1 , lowercase__ ): if random.random() < probability: graph[i].append(lowercase__ ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(lowercase__ ) return graph def _snake_case ( lowercase__ ): return { i: [j for j in range(lowercase__ ) if i != j] for i in range(lowercase__ ) } if __name__ == "__main__": import doctest doctest.testmod()
96
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ = {"""configuration_ibert""": ["""IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """IBertConfig""", """IBertOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ """IBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """IBertForMaskedLM""", """IBertForMultipleChoice""", """IBertForQuestionAnswering""", """IBertForSequenceClassification""", """IBertForTokenClassification""", """IBertModel""", """IBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
96
1
"""simple docstring""" import argparse import torch from transformers import RemBertConfig, RemBertModel, load_tf_weights_in_rembert from transformers.utils import logging logging.set_verbosity_info() def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): # Initialise PyTorch model _lowerCamelCase : Dict = RemBertConfig.from_json_file(lowercase__ ) print('Building PyTorch model from configuration: {}'.format(str(lowercase__ ) ) ) _lowerCamelCase : Any = RemBertModel(lowercase__ ) # Load weights from tf checkpoint load_tf_weights_in_rembert(lowercase__ , lowercase__ , lowercase__ ) # Save pytorch-model print('Save PyTorch model to {}'.format(lowercase__ ) ) torch.save(model.state_dict() , lowercase__ ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--rembert_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained RemBERT model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) lowercase__ = parser.parse_args() convert_rembert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.rembert_config_file, args.pytorch_dump_path)
96
"""simple docstring""" import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : Tuple = f'''{sampling_rate}''' _lowerCamelCase : str = '1' _lowerCamelCase : str = 'f32le' _lowerCamelCase : Union[str, Any] = [ 'ffmpeg', '-i', 'pipe:0', '-ac', ac, '-ar', ar, '-f', format_for_conversion, '-hide_banner', '-loglevel', 'quiet', 'pipe:1', ] try: with subprocess.Popen(lowercase__ , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: _lowerCamelCase : str = ffmpeg_process.communicate(lowercase__ ) except FileNotFoundError as error: raise ValueError('ffmpeg was not found but is required to load audio files from filename' ) from error _lowerCamelCase : List[Any] = output_stream[0] _lowerCamelCase : Tuple = np.frombuffer(lowercase__ , np.floataa ) if audio.shape[0] == 0: raise ValueError('Malformed soundfile' ) return audio def _snake_case ( lowercase__ , lowercase__ , lowercase__ = "f32le" , ): _lowerCamelCase : Optional[Any] = f'''{sampling_rate}''' _lowerCamelCase : List[str] = '1' if format_for_conversion == "s16le": _lowerCamelCase : List[str] = 2 elif format_for_conversion == "f32le": _lowerCamelCase : List[Any] = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) _lowerCamelCase : Dict = platform.system() if system == "Linux": _lowerCamelCase : Optional[int] = 'alsa' _lowerCamelCase : Optional[Any] = 'default' elif system == "Darwin": _lowerCamelCase : Optional[int] = 'avfoundation' _lowerCamelCase : Any = ':0' elif system == "Windows": _lowerCamelCase : Tuple = 'dshow' _lowerCamelCase : Tuple = 'default' _lowerCamelCase : Optional[int] = [ 'ffmpeg', '-f', format_, '-i', input_, '-ac', ac, '-ar', ar, '-f', format_for_conversion, '-fflags', 'nobuffer', '-hide_banner', '-loglevel', 'quiet', 'pipe:1', ] _lowerCamelCase : Tuple = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample _lowerCamelCase : List[Any] = _ffmpeg_stream(lowercase__ , lowercase__ ) for item in iterator: yield item def _snake_case ( lowercase__ , lowercase__ , lowercase__ = None , lowercase__ = None , lowercase__ = "f32le" , ): if stream_chunk_s is not None: _lowerCamelCase : int = stream_chunk_s else: _lowerCamelCase : Optional[Any] = chunk_length_s _lowerCamelCase : Optional[Any] = ffmpeg_microphone(lowercase__ , lowercase__ , format_for_conversion=lowercase__ ) if format_for_conversion == "s16le": _lowerCamelCase : List[str] = np.intaa _lowerCamelCase : str = 2 elif format_for_conversion == "f32le": _lowerCamelCase : Any = np.floataa _lowerCamelCase : List[Any] = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) if stride_length_s is None: _lowerCamelCase : Union[str, Any] = chunk_length_s / 6 _lowerCamelCase : Optional[int] = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(lowercase__ , (int, float) ): _lowerCamelCase : Any = [stride_length_s, stride_length_s] _lowerCamelCase : Tuple = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample _lowerCamelCase : Optional[Any] = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample _lowerCamelCase : List[Any] = datetime.datetime.now() _lowerCamelCase : Optional[int] = datetime.timedelta(seconds=lowercase__ ) for item in chunk_bytes_iter(lowercase__ , lowercase__ , stride=(stride_left, stride_right) , stream=lowercase__ ): # Put everything back in numpy scale _lowerCamelCase : List[Any] = np.frombuffer(item['raw'] , dtype=lowercase__ ) _lowerCamelCase : int = ( item['stride'][0] // size_of_sample, item['stride'][1] // size_of_sample, ) _lowerCamelCase : Optional[int] = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ = False ): _lowerCamelCase : int = B'' _lowerCamelCase, _lowerCamelCase : Dict = stride if stride_left + stride_right >= chunk_len: raise ValueError( f'''Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}''' ) _lowerCamelCase : str = 0 for raw in iterator: acc += raw if stream and len(lowercase__ ) < chunk_len: _lowerCamelCase : Optional[int] = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(lowercase__ ) >= chunk_len: # We are flushing the accumulator _lowerCamelCase : str = (_stride_left, stride_right) _lowerCamelCase : str = {'raw': acc[:chunk_len], 'stride': stride} if stream: _lowerCamelCase : List[Any] = False yield item _lowerCamelCase : Optional[Any] = stride_left _lowerCamelCase : str = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(lowercase__ ) > stride_left: _lowerCamelCase : Optional[Any] = {'raw': acc, 'stride': (_stride_left, 0)} if stream: _lowerCamelCase : Tuple = False yield item def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : int = 2**24 # 16Mo try: with subprocess.Popen(lowercase__ , stdout=subprocess.PIPE , bufsize=lowercase__ ) as ffmpeg_process: while True: _lowerCamelCase : Optional[Any] = ffmpeg_process.stdout.read(lowercase__ ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError('ffmpeg was not found but is required to stream audio files from filename' ) from error
96
1
"""simple docstring""" import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = ["""image_processor""", """tokenizer"""] lowerCamelCase__ = """ViTImageProcessor""" lowerCamelCase__ = ("""CLIPTokenizer""", """CLIPTokenizerFast""") def __init__( self , lowercase=None , lowercase=None , **lowercase ): _lowerCamelCase : Any = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , lowercase , ) _lowerCamelCase : Any = kwargs.pop('feature_extractor' ) _lowerCamelCase : int = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(lowercase , lowercase ) def __call__( self , lowercase=None , lowercase=None , lowercase=None , lowercase=None , **lowercase ): if text is None and visual_prompt is None and images is None: raise ValueError('You have to specify either text, visual prompt or images.' ) if text is not None and visual_prompt is not None: raise ValueError('You have to specify exactly one type of prompt. Either text or visual prompt.' ) if text is not None: _lowerCamelCase : Union[str, Any] = self.tokenizer(lowercase , return_tensors=lowercase , **lowercase ) if visual_prompt is not None: _lowerCamelCase : Optional[int] = self.image_processor(lowercase , return_tensors=lowercase , **lowercase ) if images is not None: _lowerCamelCase : Tuple = self.image_processor(lowercase , return_tensors=lowercase , **lowercase ) if visual_prompt is not None and images is not None: _lowerCamelCase : Optional[Any] = { 'pixel_values': image_features.pixel_values, 'conditional_pixel_values': prompt_features.pixel_values, } return encoding elif text is not None and images is not None: _lowerCamelCase : Tuple = image_features.pixel_values return encoding elif text is not None: return encoding elif visual_prompt is not None: _lowerCamelCase : List[Any] = { 'conditional_pixel_values': prompt_features.pixel_values, } return encoding else: return BatchEncoding(data=dict(**lowercase ) , tensor_type=lowercase ) def A_ ( self , *lowercase , **lowercase ): return self.tokenizer.batch_decode(*lowercase , **lowercase ) def A_ ( self , *lowercase , **lowercase ): return self.tokenizer.decode(*lowercase , **lowercase ) @property def A_ ( self ): warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , lowercase , ) return self.image_processor_class @property def A_ ( self ): warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , lowercase , ) return self.image_processor
96
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = {"""ctrl""": """https://huggingface.co/ctrl/resolve/main/config.json"""} class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """ctrl""" lowerCamelCase__ = ["""past_key_values"""] lowerCamelCase__ = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowercase=246534 , lowercase=256 , lowercase=1280 , lowercase=8192 , lowercase=48 , lowercase=16 , lowercase=0.1 , lowercase=0.1 , lowercase=1E-6 , lowercase=0.02 , lowercase=True , **lowercase , ): _lowerCamelCase : Any = vocab_size _lowerCamelCase : Dict = n_positions _lowerCamelCase : Optional[int] = n_embd _lowerCamelCase : str = n_layer _lowerCamelCase : Union[str, Any] = n_head _lowerCamelCase : Any = dff _lowerCamelCase : int = resid_pdrop _lowerCamelCase : Dict = embd_pdrop _lowerCamelCase : Union[str, Any] = layer_norm_epsilon _lowerCamelCase : Tuple = initializer_range _lowerCamelCase : str = use_cache super().__init__(**lowercase )
96
1
"""simple docstring""" import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler lowercase__ = 16 lowercase__ = 32 def _snake_case ( lowercase__ , lowercase__ = 16 , lowercase__ = "bert-base-cased" ): _lowerCamelCase : List[Any] = AutoTokenizer.from_pretrained(lowercase__ ) _lowerCamelCase : Tuple = load_dataset('glue' , 'mrpc' ) def tokenize_function(lowercase__ ): # max_length=None => use the model max length (it's actually the default) _lowerCamelCase : Union[str, Any] = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=lowercase__ , max_length=lowercase__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset _lowerCamelCase : int = datasets.map( lowercase__ , batched=lowercase__ , remove_columns=['idx', 'sentence1', 'sentence2'] , load_from_cache_file=lowercase__ ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _lowerCamelCase : Optional[int] = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(lowercase__ ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(lowercase__ , padding='max_length' , max_length=128 , return_tensors='pt' ) return tokenizer.pad(lowercase__ , padding='longest' , return_tensors='pt' ) # Instantiate dataloaders. _lowerCamelCase : List[str] = DataLoader( tokenized_datasets['train'] , shuffle=lowercase__ , collate_fn=lowercase__ , batch_size=lowercase__ ) _lowerCamelCase : int = DataLoader( tokenized_datasets['validation'] , shuffle=lowercase__ , collate_fn=lowercase__ , batch_size=lowercase__ ) return train_dataloader, eval_dataloader def _snake_case ( lowercase__ , lowercase__ ): # Initialize accelerator _lowerCamelCase : Optional[int] = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _lowerCamelCase : Optional[int] = config['lr'] _lowerCamelCase : Optional[int] = int(config['num_epochs'] ) _lowerCamelCase : Union[str, Any] = int(config['seed'] ) _lowerCamelCase : Optional[int] = int(config['batch_size'] ) _lowerCamelCase : Dict = args.model_name_or_path set_seed(lowercase__ ) _lowerCamelCase, _lowerCamelCase : Optional[int] = get_dataloaders(lowercase__ , lowercase__ , lowercase__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _lowerCamelCase : int = AutoModelForSequenceClassification.from_pretrained(lowercase__ , return_dict=lowercase__ ) # Instantiate optimizer _lowerCamelCase : Optional[int] = ( AdamW if accelerator.state.deepspeed_plugin is None or 'optimizer' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) _lowerCamelCase : Union[str, Any] = optimizer_cls(params=model.parameters() , lr=lowercase__ ) if accelerator.state.deepspeed_plugin is not None: _lowerCamelCase : str = accelerator.state.deepspeed_plugin.deepspeed_config[ 'gradient_accumulation_steps' ] else: _lowerCamelCase : Tuple = 1 _lowerCamelCase : List[Any] = (len(lowercase__ ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): _lowerCamelCase : Tuple = get_linear_schedule_with_warmup( optimizer=lowercase__ , num_warmup_steps=0 , num_training_steps=lowercase__ , ) else: _lowerCamelCase : Any = DummyScheduler(lowercase__ , total_num_steps=lowercase__ , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : Union[str, Any] = accelerator.prepare( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) # We need to keep track of how many total steps we have iterated over _lowerCamelCase : Union[str, Any] = 0 # We also need to keep track of the stating epoch so files are named properly _lowerCamelCase : Dict = 0 # Now we train the model _lowerCamelCase : Dict = evaluate.load('glue' , 'mrpc' ) _lowerCamelCase : Optional[int] = 0 _lowerCamelCase : str = {} for epoch in range(lowercase__ , lowercase__ ): model.train() for step, batch in enumerate(lowercase__ ): _lowerCamelCase : List[Any] = model(**lowercase__ ) _lowerCamelCase : int = outputs.loss _lowerCamelCase : Dict = loss / gradient_accumulation_steps accelerator.backward(lowercase__ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 model.eval() _lowerCamelCase : Union[str, Any] = 0 for step, batch in enumerate(lowercase__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _lowerCamelCase : Optional[int] = model(**lowercase__ ) _lowerCamelCase : Dict = outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times _lowerCamelCase, _lowerCamelCase : List[str] = accelerator.gather( (predictions, batch['labels']) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(lowercase__ ) - 1: _lowerCamelCase : Optional[Any] = predictions[: len(eval_dataloader.dataset ) - samples_seen] _lowerCamelCase : Dict = references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=lowercase__ , references=lowercase__ , ) _lowerCamelCase : List[Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f'''epoch {epoch}:''' , lowercase__ ) _lowerCamelCase : Tuple = eval_metric['accuracy'] if best_performance < eval_metric["accuracy"]: _lowerCamelCase : str = eval_metric['accuracy'] if args.performance_lower_bound is not None: assert ( args.performance_lower_bound <= best_performance ), f'''Best performance metric {best_performance} is lower than the lower bound {args.performance_lower_bound}''' accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , 'all_results.json' ) , 'w' ) as f: json.dump(lowercase__ , lowercase__ ) def _snake_case ( ): _lowerCamelCase : Any = argparse.ArgumentParser(description='Simple example of training script tracking peak GPU memory usage.' ) parser.add_argument( '--model_name_or_path' , type=lowercase__ , default='bert-base-cased' , help='Path to pretrained model or model identifier from huggingface.co/models.' , required=lowercase__ , ) parser.add_argument( '--output_dir' , type=lowercase__ , default='.' , help='Optional save directory where all checkpoint folders will be stored. Default is the current working directory.' , ) parser.add_argument( '--performance_lower_bound' , type=lowercase__ , default=lowercase__ , help='Optional lower bound for the performance metric. If set, the training will throw error when the performance metric drops below this value.' , ) parser.add_argument( '--num_epochs' , type=lowercase__ , default=3 , help='Number of train epochs.' , ) _lowerCamelCase : Optional[Any] = parser.parse_args() _lowerCamelCase : str = {'lr': 2E-5, 'num_epochs': args.num_epochs, 'seed': 42, 'batch_size': 16} training_function(lowercase__ , lowercase__ ) if __name__ == "__main__": main()
96
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator from typing import Any class lowerCAmelCase__ : '''simple docstring''' def __init__( self , lowercase ): _lowerCamelCase : Any = data _lowerCamelCase : Node | None = None class lowerCAmelCase__ : '''simple docstring''' def __init__( self ): _lowerCamelCase : str = None _lowerCamelCase : str = None def __iter__( self ): _lowerCamelCase : List[str] = self.head while self.head: yield node.data _lowerCamelCase : Optional[int] = node.next if node == self.head: break def __len__( self ): return sum(1 for _ in self ) def __repr__( self ): return "->".join(str(lowercase ) for item in iter(self ) ) def A_ ( self , lowercase ): self.insert_nth(len(self ) , lowercase ) def A_ ( self , lowercase ): self.insert_nth(0 , lowercase ) def A_ ( self , lowercase , lowercase ): if index < 0 or index > len(self ): raise IndexError('list index out of range.' ) _lowerCamelCase : List[Any] = Node(lowercase ) if self.head is None: _lowerCamelCase : str = new_node # first node points itself _lowerCamelCase : Union[str, Any] = new_node elif index == 0: # insert at head _lowerCamelCase : List[str] = self.head _lowerCamelCase : str = new_node else: _lowerCamelCase : Union[str, Any] = self.head for _ in range(index - 1 ): _lowerCamelCase : List[Any] = temp.next _lowerCamelCase : Union[str, Any] = temp.next _lowerCamelCase : List[str] = new_node if index == len(self ) - 1: # insert at tail _lowerCamelCase : Any = new_node def A_ ( self ): return self.delete_nth(0 ) def A_ ( self ): return self.delete_nth(len(self ) - 1 ) def A_ ( self , lowercase = 0 ): if not 0 <= index < len(self ): raise IndexError('list index out of range.' ) _lowerCamelCase : Any = self.head if self.head == self.tail: # just one node _lowerCamelCase : List[str] = None elif index == 0: # delete head node _lowerCamelCase : List[str] = self.tail.next.next _lowerCamelCase : Optional[int] = self.head.next else: _lowerCamelCase : Dict = self.head for _ in range(index - 1 ): _lowerCamelCase : List[Any] = temp.next _lowerCamelCase : int = temp.next _lowerCamelCase : Optional[int] = temp.next.next if index == len(self ) - 1: # delete at tail _lowerCamelCase : List[Any] = temp return delete_node.data def A_ ( self ): return len(self ) == 0 def _snake_case ( ): _lowerCamelCase : Union[str, Any] = CircularLinkedList() assert len(lowercase__ ) == 0 assert circular_linked_list.is_empty() is True assert str(lowercase__ ) == "" try: circular_linked_list.delete_front() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_tail() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_nth(-1 ) raise AssertionError except IndexError: assert True try: circular_linked_list.delete_nth(0 ) raise AssertionError except IndexError: assert True assert circular_linked_list.is_empty() is True for i in range(5 ): assert len(lowercase__ ) == i circular_linked_list.insert_nth(lowercase__ , i + 1 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 6 ) ) circular_linked_list.insert_tail(6 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 7 ) ) circular_linked_list.insert_head(0 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(0 , 7 ) ) assert circular_linked_list.delete_front() == 0 assert circular_linked_list.delete_tail() == 6 assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 6 ) ) assert circular_linked_list.delete_nth(2 ) == 3 circular_linked_list.insert_nth(2 , 3 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 6 ) ) assert circular_linked_list.is_empty() is False if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters lowercase__ = (720, 1280) # Height, Width lowercase__ = (0.4, 0.6) # if height or width lower than this scale, drop it. lowercase__ = 1 / 100 lowercase__ = """""" lowercase__ = """""" lowercase__ = """""" lowercase__ = 250 def _snake_case ( ): _lowerCamelCase, _lowerCamelCase : str = get_dataset(lowercase__ , lowercase__ ) for index in range(lowercase__ ): _lowerCamelCase : List[str] = random.sample(range(len(lowercase__ ) ) , 4 ) _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : str = update_image_and_anno( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , filter_scale=lowercase__ , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' _lowerCamelCase : int = random_chars(32 ) _lowerCamelCase : Dict = path.split(os.sep )[-1].rsplit('.' , 1 )[0] _lowerCamelCase : List[str] = f'''{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}''' cva.imwrite(f'''{file_root}.jpg''' , lowercase__ , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(f'''Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}''' ) _lowerCamelCase : Optional[int] = [] for anno in new_annos: _lowerCamelCase : Union[str, Any] = anno[3] - anno[1] _lowerCamelCase : Tuple = anno[4] - anno[2] _lowerCamelCase : Union[str, Any] = anno[1] + width / 2 _lowerCamelCase : Optional[int] = anno[2] + height / 2 _lowerCamelCase : int = f'''{anno[0]} {x_center} {y_center} {width} {height}''' annos_list.append(lowercase__ ) with open(f'''{file_root}.txt''' , 'w' ) as outfile: outfile.write('\n'.join(line for line in annos_list ) ) def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : Dict = [] _lowerCamelCase : int = [] for label_file in glob.glob(os.path.join(lowercase__ , '*.txt' ) ): _lowerCamelCase : List[Any] = label_file.split(os.sep )[-1].rsplit('.' , 1 )[0] with open(lowercase__ ) as in_file: _lowerCamelCase : int = in_file.readlines() _lowerCamelCase : Union[str, Any] = os.path.join(lowercase__ , f'''{label_name}.jpg''' ) _lowerCamelCase : List[Any] = [] for obj_list in obj_lists: _lowerCamelCase : int = obj_list.rstrip('\n' ).split(' ' ) _lowerCamelCase : List[str] = float(obj[1] ) - float(obj[3] ) / 2 _lowerCamelCase : str = float(obj[2] ) - float(obj[4] ) / 2 _lowerCamelCase : str = float(obj[1] ) + float(obj[3] ) / 2 _lowerCamelCase : Optional[int] = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(lowercase__ ) labels.append(lowercase__ ) return img_paths, labels def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ = 0.0 , ): _lowerCamelCase : Optional[Any] = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) _lowerCamelCase : Optional[int] = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) _lowerCamelCase : Tuple = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) _lowerCamelCase : List[str] = int(scale_x * output_size[1] ) _lowerCamelCase : str = int(scale_y * output_size[0] ) _lowerCamelCase : Optional[Any] = [] _lowerCamelCase : Tuple = [] for i, index in enumerate(lowercase__ ): _lowerCamelCase : Optional[Any] = all_img_list[index] path_list.append(lowercase__ ) _lowerCamelCase : str = all_annos[index] _lowerCamelCase : Union[str, Any] = cva.imread(lowercase__ ) if i == 0: # top-left _lowerCamelCase : Dict = cva.resize(lowercase__ , (divid_point_x, divid_point_y) ) _lowerCamelCase : Dict = img for bbox in img_annos: _lowerCamelCase : Optional[int] = bbox[1] * scale_x _lowerCamelCase : str = bbox[2] * scale_y _lowerCamelCase : List[str] = bbox[3] * scale_x _lowerCamelCase : int = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right _lowerCamelCase : int = cva.resize(lowercase__ , (output_size[1] - divid_point_x, divid_point_y) ) _lowerCamelCase : List[Any] = img for bbox in img_annos: _lowerCamelCase : Union[str, Any] = scale_x + bbox[1] * (1 - scale_x) _lowerCamelCase : List[str] = bbox[2] * scale_y _lowerCamelCase : int = scale_x + bbox[3] * (1 - scale_x) _lowerCamelCase : Optional[int] = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left _lowerCamelCase : Optional[Any] = cva.resize(lowercase__ , (divid_point_x, output_size[0] - divid_point_y) ) _lowerCamelCase : int = img for bbox in img_annos: _lowerCamelCase : Optional[int] = bbox[1] * scale_x _lowerCamelCase : List[Any] = scale_y + bbox[2] * (1 - scale_y) _lowerCamelCase : Optional[int] = bbox[3] * scale_x _lowerCamelCase : Any = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right _lowerCamelCase : Tuple = cva.resize( lowercase__ , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) _lowerCamelCase : Dict = img for bbox in img_annos: _lowerCamelCase : str = scale_x + bbox[1] * (1 - scale_x) _lowerCamelCase : Tuple = scale_y + bbox[2] * (1 - scale_y) _lowerCamelCase : List[str] = scale_x + bbox[3] * (1 - scale_x) _lowerCamelCase : Dict = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: _lowerCamelCase : List[Any] = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def _snake_case ( lowercase__ ): assert number_char > 1, "The number of character should greater than 1" _lowerCamelCase : str = ascii_lowercase + digits return "".join(random.choice(lowercase__ ) for _ in range(lowercase__ ) ) if __name__ == "__main__": main() print("""DONE ✅""")
96
"""simple docstring""" # Lint as: python3 import os import re import urllib.parse from pathlib import Path from typing import Callable, List, Optional, Union from zipfile import ZipFile from ..utils.file_utils import cached_path, hf_github_url from ..utils.logging import get_logger from ..utils.version import Version lowercase__ = get_logger(__name__) class lowerCAmelCase__ : '''simple docstring''' lowerCamelCase__ = """dummy_data""" lowerCamelCase__ = """datasets""" lowerCamelCase__ = False def __init__( self , lowercase , lowercase , lowercase , lowercase = None , lowercase = False , lowercase = True , lowercase = None , ): _lowerCamelCase : Optional[Any] = 0 _lowerCamelCase : Dict = dataset_name _lowerCamelCase : Union[str, Any] = cache_dir _lowerCamelCase : Dict = use_local_dummy_data _lowerCamelCase : Tuple = config # download_callbacks take a single url as input _lowerCamelCase : List[Callable] = download_callbacks or [] # if False, it doesn't load existing files and it returns the paths of the dummy files relative # to the dummy_data zip file root _lowerCamelCase : Any = load_existing_dummy_data # TODO(PVP, QL) might need to make this more general _lowerCamelCase : str = str(lowercase ) # to be downloaded _lowerCamelCase : Union[str, Any] = None _lowerCamelCase : int = None @property def A_ ( self ): if self._dummy_file is None: _lowerCamelCase : Tuple = self.download_dummy_data() return self._dummy_file @property def A_ ( self ): if self.config is not None: # structure is dummy / config_name / version_name return os.path.join('dummy' , self.config.name , self.version_name ) # structure is dummy / version_name return os.path.join('dummy' , self.version_name ) @property def A_ ( self ): return os.path.join(self.dummy_data_folder , 'dummy_data.zip' ) def A_ ( self ): _lowerCamelCase : List[str] = ( self.local_path_to_dummy_data if self.use_local_dummy_data is True else self.github_path_to_dummy_data ) _lowerCamelCase : int = cached_path( lowercase , cache_dir=self.cache_dir , extract_compressed_file=lowercase , force_extract=lowercase ) return os.path.join(lowercase , self.dummy_file_name ) @property def A_ ( self ): return os.path.join(self.datasets_scripts_dir , self.dataset_name , self.dummy_zip_file ) @property def A_ ( self ): if self._bucket_url is None: _lowerCamelCase : List[Any] = hf_github_url(self.dataset_name , self.dummy_zip_file.replace(os.sep , '/' ) ) return self._bucket_url @property def A_ ( self ): # return full path if its a dir if os.path.isdir(self.dummy_file ): return self.dummy_file # else cut off path to file -> example `xsum`. return "/".join(self.dummy_file.replace(os.sep , '/' ).split('/' )[:-1] ) def A_ ( self , lowercase , *lowercase ): if self.load_existing_dummy_data: # dummy data is downloaded and tested _lowerCamelCase : Union[str, Any] = self.dummy_file else: # dummy data cannot be downloaded and only the path to dummy file is returned _lowerCamelCase : Union[str, Any] = self.dummy_file_name # special case when data_url is a dict if isinstance(lowercase , lowercase ): return self.create_dummy_data_dict(lowercase , lowercase ) elif isinstance(lowercase , (list, tuple) ): return self.create_dummy_data_list(lowercase , lowercase ) else: return self.create_dummy_data_single(lowercase , lowercase ) def A_ ( self , lowercase , *lowercase ): return self.download_and_extract(lowercase ) def A_ ( self , lowercase , lowercase ): return self.download_and_extract(lowercase ) def A_ ( self , lowercase , *lowercase , **lowercase ): return path def A_ ( self ): return {} def A_ ( self , lowercase , lowercase ): _lowerCamelCase : Optional[int] = {} for key, single_urls in data_url.items(): for download_callback in self.download_callbacks: if isinstance(lowercase , lowercase ): for single_url in single_urls: download_callback(lowercase ) else: _lowerCamelCase : List[Any] = single_urls download_callback(lowercase ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus if isinstance(lowercase , lowercase ): _lowerCamelCase : List[Any] = [os.path.join(lowercase , urllib.parse.quote_plus(Path(lowercase ).name ) ) for x in single_urls] else: _lowerCamelCase : Optional[int] = single_urls _lowerCamelCase : List[Any] = os.path.join(lowercase , urllib.parse.quote_plus(Path(lowercase ).name ) ) _lowerCamelCase : int = value # make sure that values are unique if all(isinstance(lowercase , lowercase ) for i in dummy_data_dict.values() ) and len(set(dummy_data_dict.values() ) ) < len( dummy_data_dict.values() ): # append key to value to make its name unique _lowerCamelCase : List[Any] = {key: value + key for key, value in dummy_data_dict.items()} return dummy_data_dict def A_ ( self , lowercase , lowercase ): _lowerCamelCase : Optional[Any] = [] # trick: if there are many shards named like `data.txt-000001-of-00300`, only use the first one _lowerCamelCase : List[str] = all(bool(re.findall('[0-9]{3,}-of-[0-9]{3,}' , lowercase ) ) for url in data_url ) _lowerCamelCase : int = all( url.startswith('https://ftp.ncbi.nlm.nih.gov/pubmed/baseline/pubmed' ) for url in data_url ) if data_url and (is_tf_records or is_pubmed_records): _lowerCamelCase : List[str] = [data_url[0]] * len(lowercase ) for single_url in data_url: for download_callback in self.download_callbacks: download_callback(lowercase ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus _lowerCamelCase : str = os.path.join(lowercase , urllib.parse.quote_plus(single_url.split('/' )[-1] ) ) dummy_data_list.append(lowercase ) return dummy_data_list def A_ ( self , lowercase , lowercase ): for download_callback in self.download_callbacks: download_callback(lowercase ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus _lowerCamelCase : Tuple = os.path.join(lowercase , urllib.parse.quote_plus(data_url.split('/' )[-1] ) ) if os.path.exists(lowercase ) or not self.load_existing_dummy_data: return value else: # Backward compatibility, maybe deprecate at one point. # For many datasets with single url calls to dl_manager.download_and_extract, # the dummy_data.zip file is actually the zipped downloaded file # while now we expected the dummy_data.zip file to be a directory containing # the downloaded file. return path_to_dummy_data def A_ ( self ): pass def A_ ( self ): pass def A_ ( self , lowercase ): def _iter_archive_members(lowercase ): # this preserves the order of the members inside the ZIP archive _lowerCamelCase : str = Path(self.dummy_file ).parent _lowerCamelCase : Union[str, Any] = path.relative_to(lowercase ) with ZipFile(self.local_path_to_dummy_data ) as zip_file: _lowerCamelCase : List[str] = zip_file.namelist() for member in members: if member.startswith(relative_path.as_posix() ): yield dummy_parent_path.joinpath(lowercase ) _lowerCamelCase : Optional[int] = Path(lowercase ) _lowerCamelCase : Dict = _iter_archive_members(lowercase ) if self.use_local_dummy_data else path.rglob('*' ) for file_path in file_paths: if file_path.is_file() and not file_path.name.startswith(('.', '__') ): yield file_path.relative_to(lowercase ).as_posix(), file_path.open('rb' ) def A_ ( self , lowercase ): if not isinstance(lowercase , lowercase ): _lowerCamelCase : List[str] = [paths] for path in paths: if os.path.isfile(lowercase ): if os.path.basename(lowercase ).startswith(('.', '__') ): return yield path else: for dirpath, dirnames, filenames in os.walk(lowercase ): if os.path.basename(lowercase ).startswith(('.', '__') ): continue dirnames.sort() for filename in sorted(lowercase ): if filename.startswith(('.', '__') ): continue yield os.path.join(lowercase , lowercase )
96
1
"""simple docstring""" import argparse import torch from huggingface_hub import hf_hub_download from transformers import AutoTokenizer, RobertaPreLayerNormConfig, RobertaPreLayerNormForMaskedLM from transformers.utils import logging logging.set_verbosity_info() lowercase__ = logging.get_logger(__name__) def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : List[Any] = RobertaPreLayerNormConfig.from_pretrained( lowercase__ , architectures=['RobertaPreLayerNormForMaskedLM'] ) # convert state_dict _lowerCamelCase : str = torch.load(hf_hub_download(repo_id=lowercase__ , filename='pytorch_model.bin' ) ) _lowerCamelCase : Dict = {} for tensor_key, tensor_value in original_state_dict.items(): # The transformer implementation gives the model a unique name, rather than overwiriting 'roberta' if tensor_key.startswith('roberta.' ): _lowerCamelCase : Any = 'roberta_prelayernorm.' + tensor_key[len('roberta.' ) :] # The original implementation contains weights which are not used, remove them from the state_dict if tensor_key.endswith('.self.LayerNorm.weight' ) or tensor_key.endswith('.self.LayerNorm.bias' ): continue _lowerCamelCase : Dict = tensor_value _lowerCamelCase : List[Any] = RobertaPreLayerNormForMaskedLM.from_pretrained( pretrained_model_name_or_path=lowercase__ , config=lowercase__ , state_dict=lowercase__ ) model.save_pretrained(lowercase__ ) # convert tokenizer _lowerCamelCase : Dict = AutoTokenizer.from_pretrained(lowercase__ ) tokenizer.save_pretrained(lowercase__ ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--checkpoint-repo""", default=None, type=str, required=True, help="""Path the official PyTorch dump, e.g. 'andreasmadsen/efficient_mlm_m0.40'.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) lowercase__ = parser.parse_args() convert_roberta_prelayernorm_checkpoint_to_pytorch(args.checkpoint_repo, args.pytorch_dump_folder_path)
96
"""simple docstring""" def _snake_case ( lowercase__ ): stooge(lowercase__ , 0 , len(lowercase__ ) - 1 ) return arr def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: _lowerCamelCase, _lowerCamelCase : Optional[Any] = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: _lowerCamelCase : Union[str, Any] = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(lowercase__ , lowercase__ , (h - t) ) # Recursively sort last 2/3 elements stooge(lowercase__ , i + t , (lowercase__) ) # Recursively sort first 2/3 elements stooge(lowercase__ , lowercase__ , (h - t) ) if __name__ == "__main__": lowercase__ = input("""Enter numbers separated by a comma:\n""").strip() lowercase__ = [int(item) for item in user_input.split(""",""")] print(stooge_sort(unsorted))
96
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowercase__ = { """configuration_graphormer""": ["""GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GraphormerConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ """GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """GraphormerForGraphClassification""", """GraphormerModel""", """GraphormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_graphormer import GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, GraphormerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_graphormer import ( GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST, GraphormerForGraphClassification, GraphormerModel, GraphormerPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
96
"""simple docstring""" import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = ["""image_processor""", """tokenizer"""] lowerCamelCase__ = """BlipImageProcessor""" lowerCamelCase__ = """AutoTokenizer""" def __init__( self , lowercase , lowercase , lowercase ): super().__init__(lowercase , lowercase ) # add QFormer tokenizer _lowerCamelCase : int = qformer_tokenizer def __call__( self , lowercase = None , lowercase = None , lowercase = True , lowercase = False , lowercase = None , lowercase = None , lowercase = 0 , lowercase = None , lowercase = None , lowercase = False , lowercase = False , lowercase = False , lowercase = False , lowercase = False , lowercase = True , lowercase = None , **lowercase , ): if images is None and text is None: raise ValueError('You have to specify at least images or text.' ) _lowerCamelCase : int = BatchFeature() if text is not None: _lowerCamelCase : List[str] = self.tokenizer( text=lowercase , add_special_tokens=lowercase , padding=lowercase , truncation=lowercase , max_length=lowercase , stride=lowercase , pad_to_multiple_of=lowercase , return_attention_mask=lowercase , return_overflowing_tokens=lowercase , return_special_tokens_mask=lowercase , return_offsets_mapping=lowercase , return_token_type_ids=lowercase , return_length=lowercase , verbose=lowercase , return_tensors=lowercase , **lowercase , ) encoding.update(lowercase ) _lowerCamelCase : List[str] = self.qformer_tokenizer( text=lowercase , add_special_tokens=lowercase , padding=lowercase , truncation=lowercase , max_length=lowercase , stride=lowercase , pad_to_multiple_of=lowercase , return_attention_mask=lowercase , return_overflowing_tokens=lowercase , return_special_tokens_mask=lowercase , return_offsets_mapping=lowercase , return_token_type_ids=lowercase , return_length=lowercase , verbose=lowercase , return_tensors=lowercase , **lowercase , ) _lowerCamelCase : List[Any] = qformer_text_encoding.pop('input_ids' ) _lowerCamelCase : Tuple = qformer_text_encoding.pop('attention_mask' ) if images is not None: _lowerCamelCase : int = self.image_processor(lowercase , return_tensors=lowercase ) encoding.update(lowercase ) return encoding def A_ ( self , *lowercase , **lowercase ): return self.tokenizer.batch_decode(*lowercase , **lowercase ) def A_ ( self , *lowercase , **lowercase ): return self.tokenizer.decode(*lowercase , **lowercase ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def A_ ( self ): _lowerCamelCase : Union[str, Any] = self.tokenizer.model_input_names _lowerCamelCase : Any = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def A_ ( self , lowercase , **lowercase ): if os.path.isfile(lowercase ): raise ValueError(F'''Provided path ({save_directory}) should be a directory, not a file''' ) os.makedirs(lowercase , exist_ok=lowercase ) _lowerCamelCase : Optional[Any] = os.path.join(lowercase , 'qformer_tokenizer' ) self.qformer_tokenizer.save_pretrained(lowercase ) return super().save_pretrained(lowercase , **lowercase ) @classmethod def A_ ( cls , lowercase , **lowercase ): _lowerCamelCase : Union[str, Any] = AutoTokenizer.from_pretrained(lowercase , subfolder='qformer_tokenizer' ) _lowerCamelCase : Dict = cls._get_arguments_from_pretrained(lowercase , **lowercase ) args.append(lowercase ) return cls(*lowercase )
96
1
"""simple docstring""" import argparse import json import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import AutoImageProcessor, SwinConfig, SwinForImageClassification def _snake_case ( lowercase__ ): _lowerCamelCase : str = SwinConfig() _lowerCamelCase : str = swin_name.split('_' ) _lowerCamelCase : Union[str, Any] = name_split[1] _lowerCamelCase : str = int(name_split[4] ) _lowerCamelCase : List[Any] = int(name_split[3][-1] ) if model_size == "tiny": _lowerCamelCase : Any = 96 _lowerCamelCase : Optional[Any] = (2, 2, 6, 2) _lowerCamelCase : Dict = (3, 6, 12, 24) elif model_size == "small": _lowerCamelCase : Tuple = 96 _lowerCamelCase : Tuple = (2, 2, 18, 2) _lowerCamelCase : Tuple = (3, 6, 12, 24) elif model_size == "base": _lowerCamelCase : Tuple = 128 _lowerCamelCase : int = (2, 2, 18, 2) _lowerCamelCase : Tuple = (4, 8, 16, 32) else: _lowerCamelCase : Any = 192 _lowerCamelCase : Tuple = (2, 2, 18, 2) _lowerCamelCase : str = (6, 12, 24, 48) if "in22k" in swin_name: _lowerCamelCase : Any = 21841 else: _lowerCamelCase : str = 1000 _lowerCamelCase : List[Any] = 'huggingface/label-files' _lowerCamelCase : List[str] = 'imagenet-1k-id2label.json' _lowerCamelCase : Optional[Any] = json.load(open(hf_hub_download(lowercase__ , lowercase__ , repo_type='dataset' ) , 'r' ) ) _lowerCamelCase : int = {int(lowercase__ ): v for k, v in idalabel.items()} _lowerCamelCase : str = idalabel _lowerCamelCase : Optional[Any] = {v: k for k, v in idalabel.items()} _lowerCamelCase : List[str] = img_size _lowerCamelCase : List[Any] = num_classes _lowerCamelCase : Any = embed_dim _lowerCamelCase : Any = depths _lowerCamelCase : Dict = num_heads _lowerCamelCase : Any = window_size return config def _snake_case ( lowercase__ ): if "patch_embed.proj" in name: _lowerCamelCase : Optional[Any] = name.replace('patch_embed.proj' , 'embeddings.patch_embeddings.projection' ) if "patch_embed.norm" in name: _lowerCamelCase : Union[str, Any] = name.replace('patch_embed.norm' , 'embeddings.norm' ) if "layers" in name: _lowerCamelCase : Tuple = 'encoder.' + name if "attn.proj" in name: _lowerCamelCase : List[str] = name.replace('attn.proj' , 'attention.output.dense' ) if "attn" in name: _lowerCamelCase : str = name.replace('attn' , 'attention.self' ) if "norm1" in name: _lowerCamelCase : List[Any] = name.replace('norm1' , 'layernorm_before' ) if "norm2" in name: _lowerCamelCase : List[Any] = name.replace('norm2' , 'layernorm_after' ) if "mlp.fc1" in name: _lowerCamelCase : Optional[Any] = name.replace('mlp.fc1' , 'intermediate.dense' ) if "mlp.fc2" in name: _lowerCamelCase : str = name.replace('mlp.fc2' , 'output.dense' ) if name == "norm.weight": _lowerCamelCase : Union[str, Any] = 'layernorm.weight' if name == "norm.bias": _lowerCamelCase : Any = 'layernorm.bias' if "head" in name: _lowerCamelCase : Dict = name.replace('head' , 'classifier' ) else: _lowerCamelCase : Tuple = 'swin.' + name return name def _snake_case ( lowercase__ , lowercase__ ): for key in orig_state_dict.copy().keys(): _lowerCamelCase : Optional[int] = orig_state_dict.pop(lowercase__ ) if "mask" in key: continue elif "qkv" in key: _lowerCamelCase : Dict = key.split('.' ) _lowerCamelCase : Optional[Any] = int(key_split[1] ) _lowerCamelCase : Any = int(key_split[3] ) _lowerCamelCase : Union[str, Any] = model.swin.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: _lowerCamelCase : Union[str, Any] = val[:dim, :] _lowerCamelCase : int = val[ dim : dim * 2, : ] _lowerCamelCase : int = val[-dim:, :] else: _lowerCamelCase : Optional[int] = val[ :dim ] _lowerCamelCase : Dict = val[ dim : dim * 2 ] _lowerCamelCase : Optional[Any] = val[ -dim: ] else: _lowerCamelCase : str = val return orig_state_dict def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : List[str] = timm.create_model(lowercase__ , pretrained=lowercase__ ) timm_model.eval() _lowerCamelCase : Tuple = get_swin_config(lowercase__ ) _lowerCamelCase : Optional[int] = SwinForImageClassification(lowercase__ ) model.eval() _lowerCamelCase : int = convert_state_dict(timm_model.state_dict() , lowercase__ ) model.load_state_dict(lowercase__ ) _lowerCamelCase : Dict = 'http://images.cocodataset.org/val2017/000000039769.jpg' _lowerCamelCase : List[str] = AutoImageProcessor.from_pretrained('microsoft/{}'.format(swin_name.replace('_' , '-' ) ) ) _lowerCamelCase : Optional[int] = Image.open(requests.get(lowercase__ , stream=lowercase__ ).raw ) _lowerCamelCase : Dict = image_processor(images=lowercase__ , return_tensors='pt' ) _lowerCamelCase : Tuple = timm_model(inputs['pixel_values'] ) _lowerCamelCase : List[Any] = model(**lowercase__ ).logits assert torch.allclose(lowercase__ , lowercase__ , atol=1E-3 ) print(f'''Saving model {swin_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowercase__ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(lowercase__ ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--swin_name""", default="""swin_tiny_patch4_window7_224""", type=str, help="""Name of the Swin timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) lowercase__ = parser.parse_args() convert_swin_checkpoint(args.swin_name, args.pytorch_dump_folder_path)
96
"""simple docstring""" import json import logging import os import sys from time import time from unittest.mock import patch from transformers.testing_utils import TestCasePlus, require_torch_tpu logging.basicConfig(level=logging.DEBUG) lowercase__ = logging.getLogger() def _snake_case ( lowercase__ ): _lowerCamelCase : List[Any] = {} _lowerCamelCase : List[Any] = os.path.join(lowercase__ , 'all_results.json' ) if os.path.exists(lowercase__ ): with open(lowercase__ , 'r' ) as f: _lowerCamelCase : List[Any] = json.load(lowercase__ ) else: raise ValueError(f'''can\'t find {path}''' ) return results lowercase__ = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) @require_torch_tpu class lowerCAmelCase__ ( lowercase ): '''simple docstring''' def A_ ( self ): import xla_spawn _lowerCamelCase : List[Any] = self.get_auto_remove_tmp_dir() _lowerCamelCase : List[Any] = F''' ./examples/pytorch/text-classification/run_glue.py --num_cores=8 ./examples/pytorch/text-classification/run_glue.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --overwrite_output_dir --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --do_train --do_eval --debug tpu_metrics_debug --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --max_steps=10 --warmup_steps=2 --seed=42 --max_seq_length=128 '''.split() with patch.object(lowercase , 'argv' , lowercase ): _lowerCamelCase : Dict = time() xla_spawn.main() _lowerCamelCase : Any = time() _lowerCamelCase : Optional[int] = get_results(lowercase ) self.assertGreaterEqual(result['eval_accuracy'] , 0.75 ) # Assert that the script takes less than 500 seconds to make sure it doesn't hang. self.assertLess(end - start , 500 ) def A_ ( self ): import xla_spawn _lowerCamelCase : Tuple = '\n ./tests/test_trainer_tpu.py\n --num_cores=8\n ./tests/test_trainer_tpu.py\n '.split() with patch.object(lowercase , 'argv' , lowercase ): xla_spawn.main()
96
1
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() lowercase__ = logging.get_logger(__name__) def _snake_case ( lowercase__ , lowercase__=False ): _lowerCamelCase : Union[str, Any] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''blocks.{i}.norm1.weight''', f'''vit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''blocks.{i}.norm1.bias''', f'''vit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((f'''blocks.{i}.attn.proj.weight''', f'''vit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((f'''blocks.{i}.attn.proj.bias''', f'''vit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''blocks.{i}.norm2.weight''', f'''vit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''blocks.{i}.norm2.bias''', f'''vit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((f'''blocks.{i}.mlp.fc1.weight''', f'''vit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''blocks.{i}.mlp.fc1.bias''', f'''vit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''blocks.{i}.mlp.fc2.weight''', f'''vit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''blocks.{i}.mlp.fc2.bias''', f'''vit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ('cls_token', 'vit.embeddings.cls_token'), ('patch_embed.proj.weight', 'vit.embeddings.patch_embeddings.projection.weight'), ('patch_embed.proj.bias', 'vit.embeddings.patch_embeddings.projection.bias'), ('pos_embed', 'vit.embeddings.position_embeddings'), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('norm.weight', 'layernorm.weight'), ('norm.bias', 'layernorm.bias'), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _lowerCamelCase : List[Any] = [(pair[0], pair[1][4:]) if pair[1].startswith('vit' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('norm.weight', 'vit.layernorm.weight'), ('norm.bias', 'vit.layernorm.bias'), ('head.weight', 'classifier.weight'), ('head.bias', 'classifier.bias'), ] ) return rename_keys def _snake_case ( lowercase__ , lowercase__ , lowercase__=False ): for i in range(config.num_hidden_layers ): if base_model: _lowerCamelCase : Optional[Any] = '' else: _lowerCamelCase : Any = 'vit.' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _lowerCamelCase : List[Any] = state_dict.pop(f'''blocks.{i}.attn.qkv.weight''' ) _lowerCamelCase : List[str] = state_dict.pop(f'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict _lowerCamelCase : int = in_proj_weight[ : config.hidden_size, : ] _lowerCamelCase : Optional[int] = in_proj_bias[: config.hidden_size] _lowerCamelCase : Any = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _lowerCamelCase : Any = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _lowerCamelCase : Optional[Any] = in_proj_weight[ -config.hidden_size :, : ] _lowerCamelCase : Union[str, Any] = in_proj_bias[-config.hidden_size :] def _snake_case ( lowercase__ ): _lowerCamelCase : List[Any] = ['head.weight', 'head.bias'] for k in ignore_keys: state_dict.pop(lowercase__ , lowercase__ ) def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : List[Any] = dct.pop(lowercase__ ) _lowerCamelCase : str = val def _snake_case ( ): _lowerCamelCase : Union[str, Any] = 'http://images.cocodataset.org/val2017/000000039769.jpg' _lowerCamelCase : List[Any] = Image.open(requests.get(lowercase__ , stream=lowercase__ ).raw ) return im @torch.no_grad() def _snake_case ( lowercase__ , lowercase__ , lowercase__=True ): _lowerCamelCase : Optional[int] = ViTConfig() # patch_size if model_name[-1] == "8": _lowerCamelCase : Union[str, Any] = 8 # set labels if required if not base_model: _lowerCamelCase : Tuple = 1000 _lowerCamelCase : Optional[int] = 'huggingface/label-files' _lowerCamelCase : Optional[int] = 'imagenet-1k-id2label.json' _lowerCamelCase : Optional[Any] = json.load(open(hf_hub_download(lowercase__ , lowercase__ , repo_type='dataset' ) , 'r' ) ) _lowerCamelCase : List[Any] = {int(lowercase__ ): v for k, v in idalabel.items()} _lowerCamelCase : str = idalabel _lowerCamelCase : str = {v: k for k, v in idalabel.items()} # size of the architecture if model_name in ["dino_vits8", "dino_vits16"]: _lowerCamelCase : int = 384 _lowerCamelCase : List[Any] = 1536 _lowerCamelCase : List[Any] = 12 _lowerCamelCase : Dict = 6 # load original model from torch hub _lowerCamelCase : Optional[Any] = torch.hub.load('facebookresearch/dino:main' , lowercase__ ) original_model.eval() # load state_dict of original model, remove and rename some keys _lowerCamelCase : str = original_model.state_dict() if base_model: remove_classification_head_(lowercase__ ) _lowerCamelCase : Optional[Any] = create_rename_keys(lowercase__ , base_model=lowercase__ ) for src, dest in rename_keys: rename_key(lowercase__ , lowercase__ , lowercase__ ) read_in_q_k_v(lowercase__ , lowercase__ , lowercase__ ) # load HuggingFace model if base_model: _lowerCamelCase : Union[str, Any] = ViTModel(lowercase__ , add_pooling_layer=lowercase__ ).eval() else: _lowerCamelCase : Any = ViTForImageClassification(lowercase__ ).eval() model.load_state_dict(lowercase__ ) # Check outputs on an image, prepared by ViTImageProcessor _lowerCamelCase : Tuple = ViTImageProcessor() _lowerCamelCase : int = image_processor(images=prepare_img() , return_tensors='pt' ) _lowerCamelCase : Optional[int] = encoding['pixel_values'] _lowerCamelCase : int = model(lowercase__ ) if base_model: _lowerCamelCase : str = original_model(lowercase__ ) assert torch.allclose(lowercase__ , outputs.last_hidden_state[:, 0, :] , atol=1E-1 ) else: _lowerCamelCase : List[Any] = original_model(lowercase__ ) assert logits.shape == outputs.logits.shape assert torch.allclose(lowercase__ , outputs.logits , atol=1E-3 ) Path(lowercase__ ).mkdir(exist_ok=lowercase__ ) print(f'''Saving model {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowercase__ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(lowercase__ ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""dino_vitb16""", type=str, help="""Name of the model trained with DINO you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--base_model""", action="""store_true""", help="""Whether to only convert the base model (no projection head weights).""", ) parser.set_defaults(base_model=True) lowercase__ = parser.parse_args() convert_vit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.base_model)
96
"""simple docstring""" from __future__ import annotations import math import numpy as np from numpy.linalg import norm def _snake_case ( lowercase__ , lowercase__ ): return math.sqrt(sum(pow(a - b , 2 ) for a, b in zip(lowercase__ , lowercase__ ) ) ) def _snake_case ( lowercase__ , lowercase__ ): if dataset.ndim != value_array.ndim: _lowerCamelCase : Tuple = ( 'Wrong input data\'s dimensions... ' f'''dataset : {dataset.ndim}, value_array : {value_array.ndim}''' ) raise ValueError(lowercase__ ) try: if dataset.shape[1] != value_array.shape[1]: _lowerCamelCase : Optional[int] = ( 'Wrong input data\'s shape... ' f'''dataset : {dataset.shape[1]}, value_array : {value_array.shape[1]}''' ) raise ValueError(lowercase__ ) except IndexError: if dataset.ndim != value_array.ndim: raise TypeError('Wrong shape' ) if dataset.dtype != value_array.dtype: _lowerCamelCase : int = ( 'Input data have different datatype... ' f'''dataset : {dataset.dtype}, value_array : {value_array.dtype}''' ) raise TypeError(lowercase__ ) _lowerCamelCase : Optional[int] = [] for value in value_array: _lowerCamelCase : Tuple = euclidean(lowercase__ , dataset[0] ) _lowerCamelCase : Union[str, Any] = dataset[0].tolist() for dataset_value in dataset[1:]: _lowerCamelCase : Optional[Any] = euclidean(lowercase__ , lowercase__ ) if dist > temp_dist: _lowerCamelCase : List[Any] = temp_dist _lowerCamelCase : List[str] = dataset_value.tolist() answer.append([vector, dist] ) return answer def _snake_case ( lowercase__ , lowercase__ ): return np.dot(lowercase__ , lowercase__ ) / (norm(lowercase__ ) * norm(lowercase__ )) if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" from __future__ import annotations def _snake_case ( lowercase__ ): _lowerCamelCase : List[str] = 0.0_0 _lowerCamelCase : Tuple = 0 for resistor in resistors: if resistor <= 0: _lowerCamelCase : Optional[int] = f'''Resistor at index {index} has a negative or zero value!''' raise ValueError(lowercase__ ) first_sum += 1 / float(lowercase__ ) index += 1 return 1 / first_sum def _snake_case ( lowercase__ ): _lowerCamelCase : Dict = 0.0_0 _lowerCamelCase : int = 0 for resistor in resistors: sum_r += resistor if resistor < 0: _lowerCamelCase : Any = f'''Resistor at index {index} has a negative value!''' raise ValueError(lowercase__ ) index += 1 return sum_r if __name__ == "__main__": import doctest doctest.testmod()
96
"""simple docstring""" import socket def _snake_case ( ): _lowerCamelCase : List[Any] = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) _lowerCamelCase : Union[str, Any] = socket.gethostname() _lowerCamelCase : List[Any] = 12312 sock.connect((host, port) ) sock.send(B'Hello server!' ) with open('Received_file' , 'wb' ) as out_file: print('File opened' ) print('Receiving data...' ) while True: _lowerCamelCase : int = sock.recv(1024 ) if not data: break out_file.write(lowercase__ ) print('Successfully received the file' ) sock.close() print('Connection closed' ) if __name__ == "__main__": main()
96
1
"""simple docstring""" from __future__ import annotations import copy import tempfile import unittest from transformers import CONFIG_MAPPING, AutoConfig, BertConfig, GPTaConfig, TaConfig, TapasConfig, is_tf_available from transformers.testing_utils import ( DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tensorflow_probability, require_tf, slow, ) from ..bert.test_modeling_bert import BertModelTester if is_tf_available(): from transformers import ( TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelForTableQuestionAnswering, TFAutoModelForTokenClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFFunnelBaseModel, TFFunnelModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, TFTapasForQuestionAnswering, ) from transformers.models.auto.modeling_tf_auto import ( TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_MAPPING, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.tapas.modeling_tf_tapas import TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """new-model""" if is_tf_available(): class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = NewModelConfig @require_tf class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' @slow def A_ ( self ): _lowerCamelCase : List[str] = 'bert-base-cased' _lowerCamelCase : Union[str, Any] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Union[str, Any] = TFAutoModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): _lowerCamelCase : List[str] = 'bert-base-cased' _lowerCamelCase : int = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : int = TFAutoModelForPreTraining.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Dict = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : int = TFAutoModelForCausalLM.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : str = TFAutoModelForCausalLM.from_pretrained(lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : List[Any] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : str = TFAutoModelWithLMHead.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Tuple = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Optional[int] = TFAutoModelForMaskedLM.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : Tuple = TFAutoModelForMaskedLM.from_pretrained(lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Optional[int] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Tuple = TFAutoModelForSeqaSeqLM.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : Tuple = TFAutoModelForSeqaSeqLM.from_pretrained(lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: _lowerCamelCase : str = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Union[str, Any] = TFAutoModelForSequenceClassification.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: _lowerCamelCase : Optional[Any] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : List[str] = TFAutoModelForQuestionAnswering.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow @require_tensorflow_probability def A_ ( self ): for model_name in TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST[5:6]: _lowerCamelCase : Dict = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Dict = TFAutoModelForTableQuestionAnswering.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : List[Any] = TFAutoModelForTableQuestionAnswering.from_pretrained( lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) def A_ ( self ): _lowerCamelCase : int = TFAutoModelWithLMHead.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowercase ) , 14410 ) def A_ ( self ): _lowerCamelCase : Any = TFAutoModelWithLMHead.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowercase ) , 14410 ) def A_ ( self ): # For the auto model mapping, FunnelConfig has two models: FunnelModel and FunnelBaseModel _lowerCamelCase : List[str] = TFAutoModel.from_pretrained('sgugger/funnel-random-tiny' ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Optional[int] = copy.deepcopy(model.config ) _lowerCamelCase : Dict = ['FunnelBaseModel'] _lowerCamelCase : List[Any] = TFAutoModel.from_config(lowercase ) self.assertIsInstance(lowercase , lowercase ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowercase ) _lowerCamelCase : Tuple = TFAutoModel.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) def A_ ( self ): try: AutoConfig.register('new-model' , lowercase ) _lowerCamelCase : Tuple = [ TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSequenceClassification, TFAutoModelForTokenClassification, ] for auto_class in auto_classes: with self.subTest(auto_class.__name__ ): # Wrong config class will raise an error with self.assertRaises(lowercase ): auto_class.register(lowercase , lowercase ) auto_class.register(lowercase , lowercase ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(lowercase ): auto_class.register(lowercase , lowercase ) # Now that the config is registered, it can be used as any other config with the auto-API _lowerCamelCase : Optional[Any] = BertModelTester(self ).get_config() _lowerCamelCase : Dict = NewModelConfig(**tiny_config.to_dict() ) _lowerCamelCase : int = auto_class.from_config(lowercase ) self.assertIsInstance(lowercase , lowercase ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowercase ) _lowerCamelCase : List[Any] = auto_class.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"] for mapping in ( TF_MODEL_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, ): if NewModelConfig in mapping._extra_content: del mapping._extra_content[NewModelConfig] def A_ ( self ): with self.assertRaisesRegex( lowercase , 'bert-base is not a local folder and is not a valid model identifier' ): _lowerCamelCase : Union[str, Any] = TFAutoModel.from_pretrained('bert-base' ) def A_ ( self ): with self.assertRaisesRegex( lowercase , r'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): _lowerCamelCase : str = TFAutoModel.from_pretrained(lowercase , revision='aaaaaa' ) def A_ ( self ): with self.assertRaisesRegex( lowercase , 'hf-internal-testing/config-no-model does not appear to have a file named pytorch_model.bin' , ): _lowerCamelCase : Tuple = TFAutoModel.from_pretrained('hf-internal-testing/config-no-model' ) def A_ ( self ): with self.assertRaisesRegex(lowercase , 'Use `from_pt=True` to load this model' ): _lowerCamelCase : Tuple = TFAutoModel.from_pretrained('hf-internal-testing/tiny-bert-pt-only' ) def A_ ( self ): # Make sure we have cached the model. _lowerCamelCase : Optional[int] = TFAutoModel.from_pretrained('hf-internal-testing/tiny-random-bert' ) with RequestCounter() as counter: _lowerCamelCase : Optional[int] = TFAutoModel.from_pretrained('hf-internal-testing/tiny-random-bert' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 ) # With a sharded checkpoint _lowerCamelCase : int = TFAutoModel.from_pretrained('ArthurZ/tiny-random-bert-sharded' ) with RequestCounter() as counter: _lowerCamelCase : List[Any] = TFAutoModel.from_pretrained('ArthurZ/tiny-random-bert-sharded' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 )
96
"""simple docstring""" from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record lowercase__ = """\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } """ lowercase__ = """\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. """ lowercase__ = """ Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for 'record': list of question-answer dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'prediction_text': the predicted answer text - for 'multirc': list of question-answer dictionaries with the following keys: - 'idx': index of the question-answer pair as specified by the dataset - 'prediction': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for 'record': list of question-answers dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'answers': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for 'record': - 'exact_match': Exact match between answer and gold answer - 'f1': F1 score - for 'multirc': - 'exact_match': Exact match between answer and gold answer - 'f1_m': Per-question macro-F1 score - 'f1_a': Average F1 score over all answers - for 'axb': 'matthews_correlation': Matthew Correlation - for 'cb': - 'accuracy': Accuracy - 'f1': F1 score - for all others: - 'accuracy': Accuracy Examples: >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'cb') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'record') >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}] >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc') >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'axb') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def _snake_case ( lowercase__ , lowercase__ ): return float((preds == labels).mean() ) def _snake_case ( lowercase__ , lowercase__ , lowercase__="binary" ): _lowerCamelCase : str = simple_accuracy(lowercase__ , lowercase__ ) _lowerCamelCase : Any = float(fa_score(y_true=lowercase__ , y_pred=lowercase__ , average=lowercase__ ) ) return { "accuracy": acc, "f1": fa, } def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : Any = {} for id_pred, label in zip(lowercase__ , lowercase__ ): _lowerCamelCase : Tuple = f'''{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}''' _lowerCamelCase : Union[str, Any] = id_pred['prediction'] if question_id in question_map: question_map[question_id].append((pred, label) ) else: _lowerCamelCase : Optional[Any] = [(pred, label)] _lowerCamelCase, _lowerCamelCase : Optional[int] = [], [] for question, preds_labels in question_map.items(): _lowerCamelCase, _lowerCamelCase : Tuple = zip(*lowercase__ ) _lowerCamelCase : List[str] = fa_score(y_true=lowercase__ , y_pred=lowercase__ , average='macro' ) fas.append(lowercase__ ) _lowerCamelCase : int = int(sum(pred == label for pred, label in preds_labels ) == len(lowercase__ ) ) ems.append(lowercase__ ) _lowerCamelCase : Optional[Any] = float(sum(lowercase__ ) / len(lowercase__ ) ) _lowerCamelCase : Optional[int] = sum(lowercase__ ) / len(lowercase__ ) _lowerCamelCase : List[Any] = float(fa_score(y_true=lowercase__ , y_pred=[id_pred['prediction'] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCAmelCase__ ( datasets.Metric ): '''simple docstring''' def A_ ( self ): if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( 'You should supply a configuration name selected in ' '["boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg",]' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format='numpy' if not self.config_name == 'record' and not self.config_name == 'multirc' else None , ) def A_ ( self ): if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value('int64' ), "query": datasets.Value('int64' ), }, "prediction_text": datasets.Value('string' ), }, "references": { "idx": { "passage": datasets.Value('int64' ), "query": datasets.Value('int64' ), }, "answers": datasets.Sequence(datasets.Value('string' ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value('int64' ), "paragraph": datasets.Value('int64' ), "question": datasets.Value('int64' ), }, "prediction": datasets.Value('int64' ), }, "references": datasets.Value('int64' ), } else: return { "predictions": datasets.Value('int64' ), "references": datasets.Value('int64' ), } def A_ ( self , lowercase , lowercase ): if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(lowercase , lowercase )} elif self.config_name == "cb": return acc_and_fa(lowercase , lowercase , fa_avg='macro' ) elif self.config_name == "record": _lowerCamelCase : List[str] = [ { 'qas': [ {'id': ref['idx']['query'], 'answers': [{'text': ans} for ans in ref['answers']]} for ref in references ] } ] _lowerCamelCase : Union[str, Any] = {pred['idx']['query']: pred['prediction_text'] for pred in predictions} return evaluate_record(lowercase , lowercase )[0] elif self.config_name == "multirc": return evaluate_multirc(lowercase , lowercase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(lowercase , lowercase )} else: raise KeyError( 'You should supply a configuration name selected in ' '["boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg",]' )
96
1
"""simple docstring""" import collections import importlib.util import os import re from pathlib import Path lowercase__ = """src/transformers""" # Matches is_xxx_available() lowercase__ = re.compile(R"""is\_([a-z_]*)_available()""") # Catches a one-line _import_struct = {xxx} lowercase__ = re.compile(R"""^_import_structure\s+=\s+\{([^\}]+)\}""") # Catches a line with a key-values pattern: "bla": ["foo", "bar"] lowercase__ = re.compile(R"""\s+\"\S*\":\s+\[([^\]]*)\]""") # Catches a line if not is_foo_available lowercase__ = re.compile(R"""^\s*if\s+not\s+is\_[a-z_]*\_available\(\)""") # Catches a line _import_struct["bla"].append("foo") lowercase__ = re.compile(R"""^\s*_import_structure\[\"\S*\"\]\.append\(\"(\S*)\"\)""") # Catches a line _import_struct["bla"].extend(["foo", "bar"]) or _import_struct["bla"] = ["foo", "bar"] lowercase__ = re.compile(R"""^\s*_import_structure\[\S*\](?:\.extend\(|\s*=\s+)\[([^\]]*)\]""") # Catches a line with an object between quotes and a comma: "MyModel", lowercase__ = re.compile("""^\s+\"([^\"]+)\",""") # Catches a line with objects between brackets only: ["foo", "bar"], lowercase__ = re.compile("""^\s+\[([^\]]+)\]""") # Catches a line with from foo import bar, bla, boo lowercase__ = re.compile(R"""\s+from\s+\S*\s+import\s+([^\(\s].*)\n""") # Catches a line with try: lowercase__ = re.compile(R"""^\s*try:""") # Catches a line with else: lowercase__ = re.compile(R"""^\s*else:""") def _snake_case ( lowercase__ ): if _re_test_backend.search(lowercase__ ) is None: return None _lowerCamelCase : Optional[Any] = [b[0] for b in _re_backend.findall(lowercase__ )] backends.sort() return "_and_".join(lowercase__ ) def _snake_case ( lowercase__ ): with open(lowercase__ , 'r' , encoding='utf-8' , newline='\n' ) as f: _lowerCamelCase : Dict = f.readlines() _lowerCamelCase : Optional[Any] = 0 while line_index < len(lowercase__ ) and not lines[line_index].startswith('_import_structure = {' ): line_index += 1 # If this is a traditional init, just return. if line_index >= len(lowercase__ ): return None # First grab the objects without a specific backend in _import_structure _lowerCamelCase : str = [] while not lines[line_index].startswith('if TYPE_CHECKING' ) and find_backend(lines[line_index] ) is None: _lowerCamelCase : Optional[int] = lines[line_index] # If we have everything on a single line, let's deal with it. if _re_one_line_import_struct.search(lowercase__ ): _lowerCamelCase : Optional[Any] = _re_one_line_import_struct.search(lowercase__ ).groups()[0] _lowerCamelCase : Optional[Any] = re.findall('\[([^\]]+)\]' , lowercase__ ) for imp in imports: objects.extend([obj[1:-1] for obj in imp.split(', ' )] ) line_index += 1 continue _lowerCamelCase : int = _re_import_struct_key_value.search(lowercase__ ) if single_line_import_search is not None: _lowerCamelCase : Optional[Any] = [obj[1:-1] for obj in single_line_import_search.groups()[0].split(', ' ) if len(lowercase__ ) > 0] objects.extend(lowercase__ ) elif line.startswith(' ' * 8 + '"' ): objects.append(line[9:-3] ) line_index += 1 _lowerCamelCase : Optional[int] = {'none': objects} # Let's continue with backend-specific objects in _import_structure while not lines[line_index].startswith('if TYPE_CHECKING' ): # If the line is an if not is_backend_available, we grab all objects associated. _lowerCamelCase : Optional[int] = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: _lowerCamelCase : Union[str, Any] = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 _lowerCamelCase : Tuple = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(' ' * 4 ): _lowerCamelCase : Optional[int] = lines[line_index] if _re_import_struct_add_one.search(lowercase__ ) is not None: objects.append(_re_import_struct_add_one.search(lowercase__ ).groups()[0] ) elif _re_import_struct_add_many.search(lowercase__ ) is not None: _lowerCamelCase : Dict = _re_import_struct_add_many.search(lowercase__ ).groups()[0].split(', ' ) _lowerCamelCase : str = [obj[1:-1] for obj in imports if len(lowercase__ ) > 0] objects.extend(lowercase__ ) elif _re_between_brackets.search(lowercase__ ) is not None: _lowerCamelCase : Optional[Any] = _re_between_brackets.search(lowercase__ ).groups()[0].split(', ' ) _lowerCamelCase : Optional[Any] = [obj[1:-1] for obj in imports if len(lowercase__ ) > 0] objects.extend(lowercase__ ) elif _re_quote_object.search(lowercase__ ) is not None: objects.append(_re_quote_object.search(lowercase__ ).groups()[0] ) elif line.startswith(' ' * 8 + '"' ): objects.append(line[9:-3] ) elif line.startswith(' ' * 12 + '"' ): objects.append(line[13:-3] ) line_index += 1 _lowerCamelCase : Optional[int] = objects else: line_index += 1 # At this stage we are in the TYPE_CHECKING part, first grab the objects without a specific backend _lowerCamelCase : List[str] = [] while ( line_index < len(lowercase__ ) and find_backend(lines[line_index] ) is None and not lines[line_index].startswith('else' ) ): _lowerCamelCase : Tuple = lines[line_index] _lowerCamelCase : Optional[int] = _re_import.search(lowercase__ ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(', ' ) ) elif line.startswith(' ' * 8 ): objects.append(line[8:-2] ) line_index += 1 _lowerCamelCase : Optional[int] = {'none': objects} # Let's continue with backend-specific objects while line_index < len(lowercase__ ): # If the line is an if is_backend_available, we grab all objects associated. _lowerCamelCase : Tuple = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: _lowerCamelCase : List[Any] = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 _lowerCamelCase : Tuple = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(' ' * 8 ): _lowerCamelCase : List[str] = lines[line_index] _lowerCamelCase : List[Any] = _re_import.search(lowercase__ ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(', ' ) ) elif line.startswith(' ' * 12 ): objects.append(line[12:-2] ) line_index += 1 _lowerCamelCase : Dict = objects else: line_index += 1 return import_dict_objects, type_hint_objects def _snake_case ( lowercase__ , lowercase__ ): def find_duplicates(lowercase__ ): return [k for k, v in collections.Counter(lowercase__ ).items() if v > 1] if list(import_dict_objects.keys() ) != list(type_hint_objects.keys() ): return ["Both sides of the init do not have the same backends!"] _lowerCamelCase : Optional[Any] = [] for key in import_dict_objects.keys(): _lowerCamelCase : int = find_duplicates(import_dict_objects[key] ) if duplicate_imports: errors.append(f'''Duplicate _import_structure definitions for: {duplicate_imports}''' ) _lowerCamelCase : Any = find_duplicates(type_hint_objects[key] ) if duplicate_type_hints: errors.append(f'''Duplicate TYPE_CHECKING objects for: {duplicate_type_hints}''' ) if sorted(set(import_dict_objects[key] ) ) != sorted(set(type_hint_objects[key] ) ): _lowerCamelCase : Dict = 'base imports' if key == 'none' else f'''{key} backend''' errors.append(f'''Differences for {name}:''' ) for a in type_hint_objects[key]: if a not in import_dict_objects[key]: errors.append(f''' {a} in TYPE_HINT but not in _import_structure.''' ) for a in import_dict_objects[key]: if a not in type_hint_objects[key]: errors.append(f''' {a} in _import_structure but not in TYPE_HINT.''' ) return errors def _snake_case ( ): _lowerCamelCase : int = [] for root, _, files in os.walk(lowercase__ ): if "__init__.py" in files: _lowerCamelCase : Dict = os.path.join(lowercase__ , '__init__.py' ) _lowerCamelCase : Any = parse_init(lowercase__ ) if objects is not None: _lowerCamelCase : str = analyze_results(*lowercase__ ) if len(lowercase__ ) > 0: _lowerCamelCase : Tuple = f'''Problem in {fname}, both halves do not define the same objects.\n{errors[0]}''' failures.append('\n'.join(lowercase__ ) ) if len(lowercase__ ) > 0: raise ValueError('\n\n'.join(lowercase__ ) ) def _snake_case ( ): _lowerCamelCase : Dict = [] for path, directories, files in os.walk(lowercase__ ): for folder in directories: # Ignore private modules if folder.startswith('_' ): directories.remove(lowercase__ ) continue # Ignore leftovers from branches (empty folders apart from pycache) if len(list((Path(lowercase__ ) / folder).glob('*.py' ) ) ) == 0: continue _lowerCamelCase : Tuple = str((Path(lowercase__ ) / folder).relative_to(lowercase__ ) ) _lowerCamelCase : str = short_path.replace(os.path.sep , '.' ) submodules.append(lowercase__ ) for fname in files: if fname == "__init__.py": continue _lowerCamelCase : List[str] = str((Path(lowercase__ ) / fname).relative_to(lowercase__ ) ) _lowerCamelCase : int = short_path.replace('.py' , '' ).replace(os.path.sep , '.' ) if len(submodule.split('.' ) ) == 1: submodules.append(lowercase__ ) return submodules lowercase__ = [ """convert_pytorch_checkpoint_to_tf2""", """modeling_flax_pytorch_utils""", ] def _snake_case ( ): # This is to make sure the transformers module imported is the one in the repo. _lowerCamelCase : int = importlib.util.spec_from_file_location( 'transformers' , os.path.join(lowercase__ , '__init__.py' ) , submodule_search_locations=[PATH_TO_TRANSFORMERS] , ) _lowerCamelCase : List[str] = spec.loader.load_module() _lowerCamelCase : Dict = [ module for module in get_transformers_submodules() if module not in IGNORE_SUBMODULES and module not in transformers._import_structure.keys() ] if len(lowercase__ ) > 0: _lowerCamelCase : List[Any] = '\n'.join(f'''- {module}''' for module in module_not_registered ) raise ValueError( 'The following submodules are not properly registered in the main init of Transformers:\n' f'''{list_of_modules}\n''' 'Make sure they appear somewhere in the keys of `_import_structure` with an empty list as value.' ) if __name__ == "__main__": check_all_inits() check_submodules()
96
"""simple docstring""" import unittest import numpy as np import torch from diffusers import DDIMPipeline, DDIMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow, torch_device from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class lowerCAmelCase__ ( lowercase, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = DDIMPipeline lowerCamelCase__ = UNCONDITIONAL_IMAGE_GENERATION_PARAMS lowerCamelCase__ = PipelineTesterMixin.required_optional_params - { """num_images_per_prompt""", """latents""", """callback""", """callback_steps""", } lowerCamelCase__ = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS lowerCamelCase__ = False def A_ ( self ): torch.manual_seed(0 ) _lowerCamelCase : List[Any] = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('DownBlock2D', 'AttnDownBlock2D') , up_block_types=('AttnUpBlock2D', 'UpBlock2D') , ) _lowerCamelCase : List[str] = DDIMScheduler() _lowerCamelCase : Optional[int] = {'unet': unet, 'scheduler': scheduler} return components def A_ ( self , lowercase , lowercase=0 ): if str(lowercase ).startswith('mps' ): _lowerCamelCase : Dict = torch.manual_seed(lowercase ) else: _lowerCamelCase : List[str] = torch.Generator(device=lowercase ).manual_seed(lowercase ) _lowerCamelCase : Tuple = { 'batch_size': 1, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs def A_ ( self ): _lowerCamelCase : Any = 'cpu' _lowerCamelCase : Tuple = self.get_dummy_components() _lowerCamelCase : Optional[Any] = self.pipeline_class(**lowercase ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : str = self.get_dummy_inputs(lowercase ) _lowerCamelCase : int = pipe(**lowercase ).images _lowerCamelCase : Any = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 32, 32, 3) ) _lowerCamelCase : Tuple = np.array( [1.000E00, 5.717E-01, 4.717E-01, 1.000E00, 0.000E00, 1.000E00, 3.000E-04, 0.000E00, 9.000E-04] ) _lowerCamelCase : str = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowercase , 1E-3 ) def A_ ( self ): super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) def A_ ( self ): super().test_save_load_local(expected_max_difference=3E-3 ) def A_ ( self ): super().test_save_load_optional_components(expected_max_difference=3E-3 ) def A_ ( self ): super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def A_ ( self ): _lowerCamelCase : Optional[Any] = 'google/ddpm-cifar10-32' _lowerCamelCase : Optional[Any] = UNetaDModel.from_pretrained(lowercase ) _lowerCamelCase : Dict = DDIMScheduler() _lowerCamelCase : Dict = DDIMPipeline(unet=lowercase , scheduler=lowercase ) ddim.to(lowercase ) ddim.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : List[str] = torch.manual_seed(0 ) _lowerCamelCase : str = ddim(generator=lowercase , eta=0.0 , output_type='numpy' ).images _lowerCamelCase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _lowerCamelCase : List[Any] = np.array([0.17_23, 0.16_17, 0.16_00, 0.16_26, 0.14_97, 0.15_13, 0.15_05, 0.14_42, 0.14_53] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def A_ ( self ): _lowerCamelCase : Optional[int] = 'google/ddpm-ema-bedroom-256' _lowerCamelCase : str = UNetaDModel.from_pretrained(lowercase ) _lowerCamelCase : str = DDIMScheduler.from_pretrained(lowercase ) _lowerCamelCase : Optional[int] = DDIMPipeline(unet=lowercase , scheduler=lowercase ) ddpm.to(lowercase ) ddpm.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : Tuple = torch.manual_seed(0 ) _lowerCamelCase : int = ddpm(generator=lowercase , output_type='numpy' ).images _lowerCamelCase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _lowerCamelCase : str = np.array([0.00_60, 0.02_01, 0.03_44, 0.00_24, 0.00_18, 0.00_02, 0.00_22, 0.00_00, 0.00_69] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
96
1
"""simple docstring""" from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING lowercase__ = logging.get_logger(__name__) @add_end_docstrings(lowercase ) class lowerCAmelCase__ ( lowercase ): '''simple docstring''' def __init__( self , *lowercase , **lowercase ): super().__init__(*lowercase , **lowercase ) requires_backends(self , 'vision' ) self.check_model_type( TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING if self.framework == 'tf' else MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING ) def A_ ( self , lowercase=None ): _lowerCamelCase : List[Any] = {} if top_k is not None: _lowerCamelCase : List[Any] = top_k return {}, {}, postprocess_params def __call__( self , lowercase , **lowercase ): return super().__call__(lowercase , **lowercase ) def A_ ( self , lowercase ): _lowerCamelCase : Dict = load_image(lowercase ) _lowerCamelCase : Optional[Any] = self.image_processor(images=lowercase , return_tensors=self.framework ) return model_inputs def A_ ( self , lowercase ): _lowerCamelCase : Union[str, Any] = self.model(**lowercase ) return model_outputs def A_ ( self , lowercase , lowercase=5 ): if top_k > self.model.config.num_labels: _lowerCamelCase : List[str] = self.model.config.num_labels if self.framework == "pt": _lowerCamelCase : Union[str, Any] = model_outputs.logits.softmax(-1 )[0] _lowerCamelCase, _lowerCamelCase : str = probs.topk(lowercase ) elif self.framework == "tf": _lowerCamelCase : List[Any] = stable_softmax(model_outputs.logits , axis=-1 )[0] _lowerCamelCase : Tuple = tf.math.top_k(lowercase , k=lowercase ) _lowerCamelCase, _lowerCamelCase : str = topk.values.numpy(), topk.indices.numpy() else: raise ValueError(F'''Unsupported framework: {self.framework}''' ) _lowerCamelCase : str = scores.tolist() _lowerCamelCase : Optional[int] = ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(lowercase , lowercase )]
96
"""simple docstring""" # Imports import numpy as np class lowerCAmelCase__ : '''simple docstring''' def __init__( self , lowercase=None , lowercase=None , lowercase=None , lowercase=None , lowercase=None ): self.set_matricies(red=lowercase , green=lowercase , blue=lowercase , red_edge=lowercase , nir=lowercase ) def A_ ( self , lowercase=None , lowercase=None , lowercase=None , lowercase=None , lowercase=None ): if red is not None: _lowerCamelCase : Optional[int] = red if green is not None: _lowerCamelCase : Optional[Any] = green if blue is not None: _lowerCamelCase : Tuple = blue if red_edge is not None: _lowerCamelCase : Optional[Any] = red_edge if nir is not None: _lowerCamelCase : Union[str, Any] = nir return True def A_ ( self , lowercase="" , lowercase=None , lowercase=None , lowercase=None , lowercase=None , lowercase=None ): self.set_matricies(red=lowercase , green=lowercase , blue=lowercase , red_edge=lowercase , nir=lowercase ) _lowerCamelCase : str = { 'ARVI2': self.arvaa, 'CCCI': self.ccci, 'CVI': self.cvi, 'GLI': self.gli, 'NDVI': self.ndvi, 'BNDVI': self.bndvi, 'redEdgeNDVI': self.red_edge_ndvi, 'GNDVI': self.gndvi, 'GBNDVI': self.gbndvi, 'GRNDVI': self.grndvi, 'RBNDVI': self.rbndvi, 'PNDVI': self.pndvi, 'ATSAVI': self.atsavi, 'BWDRVI': self.bwdrvi, 'CIgreen': self.ci_green, 'CIrededge': self.ci_rededge, 'CI': self.ci, 'CTVI': self.ctvi, 'GDVI': self.gdvi, 'EVI': self.evi, 'GEMI': self.gemi, 'GOSAVI': self.gosavi, 'GSAVI': self.gsavi, 'Hue': self.hue, 'IVI': self.ivi, 'IPVI': self.ipvi, 'I': self.i, 'RVI': self.rvi, 'MRVI': self.mrvi, 'MSAVI': self.m_savi, 'NormG': self.norm_g, 'NormNIR': self.norm_nir, 'NormR': self.norm_r, 'NGRDI': self.ngrdi, 'RI': self.ri, 'S': self.s, 'IF': self._if, 'DVI': self.dvi, 'TVI': self.tvi, 'NDRE': self.ndre, } try: return funcs[index]() except KeyError: print('Index not in the list!' ) return False def A_ ( self ): return -0.18 + (1.17 * ((self.nir - self.red) / (self.nir + self.red))) def A_ ( self ): return ((self.nir - self.redEdge) / (self.nir + self.redEdge)) / ( (self.nir - self.red) / (self.nir + self.red) ) def A_ ( self ): return self.nir * (self.red / (self.green**2)) def A_ ( self ): return (2 * self.green - self.red - self.blue) / ( 2 * self.green + self.red + self.blue ) def A_ ( self ): return (self.nir - self.red) / (self.nir + self.red) def A_ ( self ): return (self.nir - self.blue) / (self.nir + self.blue) def A_ ( self ): return (self.redEdge - self.red) / (self.redEdge + self.red) def A_ ( self ): return (self.nir - self.green) / (self.nir + self.green) def A_ ( self ): return (self.nir - (self.green + self.blue)) / ( self.nir + (self.green + self.blue) ) def A_ ( self ): return (self.nir - (self.green + self.red)) / ( self.nir + (self.green + self.red) ) def A_ ( self ): return (self.nir - (self.blue + self.red)) / (self.nir + (self.blue + self.red)) def A_ ( self ): return (self.nir - (self.green + self.red + self.blue)) / ( self.nir + (self.green + self.red + self.blue) ) def A_ ( self , lowercase=0.08 , lowercase=1.22 , lowercase=0.03 ): return a * ( (self.nir - a * self.red - b) / (a * self.nir + self.red - a * b + x * (1 + a**2)) ) def A_ ( self ): return (0.1 * self.nir - self.blue) / (0.1 * self.nir + self.blue) def A_ ( self ): return (self.nir / self.green) - 1 def A_ ( self ): return (self.nir / self.redEdge) - 1 def A_ ( self ): return (self.red - self.blue) / self.red def A_ ( self ): _lowerCamelCase : Any = self.ndvi() return ((ndvi + 0.5) / (abs(ndvi + 0.5 ))) * (abs(ndvi + 0.5 ) ** (1 / 2)) def A_ ( self ): return self.nir - self.green def A_ ( self ): return 2.5 * ( (self.nir - self.red) / (self.nir + 6 * self.red - 7.5 * self.blue + 1) ) def A_ ( self ): _lowerCamelCase : Any = (2 * (self.nir**2 - self.red**2) + 1.5 * self.nir + 0.5 * self.red) / ( self.nir + self.red + 0.5 ) return n * (1 - 0.25 * n) - (self.red - 0.1_25) / (1 - self.red) def A_ ( self , lowercase=0.16 ): return (self.nir - self.green) / (self.nir + self.green + y) def A_ ( self , lowercase=0.5 ): return ((self.nir - self.green) / (self.nir + self.green + n)) * (1 + n) def A_ ( self ): return np.arctan( ((2 * self.red - self.green - self.blue) / 30.5) * (self.green - self.blue) ) def A_ ( self , lowercase=None , lowercase=None ): return (self.nir - b) / (a * self.red) def A_ ( self ): return (self.nir / ((self.nir + self.red) / 2)) * (self.ndvi() + 1) def A_ ( self ): return (self.red + self.green + self.blue) / 30.5 def A_ ( self ): return self.nir / self.red def A_ ( self ): return (self.rvi() - 1) / (self.rvi() + 1) def A_ ( self ): return ( (2 * self.nir + 1) - ((2 * self.nir + 1) ** 2 - 8 * (self.nir - self.red)) ** (1 / 2) ) / 2 def A_ ( self ): return self.green / (self.nir + self.red + self.green) def A_ ( self ): return self.nir / (self.nir + self.red + self.green) def A_ ( self ): return self.red / (self.nir + self.red + self.green) def A_ ( self ): return (self.green - self.red) / (self.green + self.red) def A_ ( self ): return (self.red - self.green) / (self.red + self.green) def A_ ( self ): _lowerCamelCase : Union[str, Any] = np.max([np.max(self.red ), np.max(self.green ), np.max(self.blue )] ) _lowerCamelCase : Dict = np.min([np.min(self.red ), np.min(self.green ), np.min(self.blue )] ) return (max_value - min_value) / max_value def A_ ( self ): return (2 * self.red - self.green - self.blue) / (self.green - self.blue) def A_ ( self ): return self.nir / self.red def A_ ( self ): return (self.ndvi() + 0.5) ** (1 / 2) def A_ ( self ): return (self.nir - self.redEdge) / (self.nir + self.redEdge)
96
1
"""simple docstring""" import os def _snake_case ( ): _lowerCamelCase : Dict = os.path.dirname(os.path.realpath(lowercase__ ) ) _lowerCamelCase : int = os.path.join(lowercase__ , 'triangle.txt' ) with open(lowercase__ ) as f: _lowerCamelCase : str = f.readlines() _lowerCamelCase : Optional[Any] = [] for line in triangle: _lowerCamelCase : Optional[int] = [] for number in line.strip().split(' ' ): numbers_from_line.append(int(lowercase__ ) ) a.append(lowercase__ ) for i in range(1 , len(lowercase__ ) ): for j in range(len(a[i] ) ): _lowerCamelCase : str = a[i - 1][j] if j != len(a[i - 1] ) else 0 _lowerCamelCase : Any = a[i - 1][j - 1] if j > 0 else 0 a[i][j] += max(lowercase__ , lowercase__ ) return max(a[-1] ) if __name__ == "__main__": print(solution())
96
"""simple docstring""" import torch from torch import nn from transformers import CLIPPreTrainedModel, CLIPVisionModel from ...models.attention import BasicTransformerBlock from ...utils import logging lowercase__ = logging.get_logger(__name__) # pylint: disable=invalid-name class lowerCAmelCase__ ( lowercase ): '''simple docstring''' def __init__( self , lowercase , lowercase=768 ): super().__init__(lowercase ) _lowerCamelCase : Any = proj_size _lowerCamelCase : Dict = CLIPVisionModel(lowercase ) _lowerCamelCase : List[str] = PaintByExampleMapper(lowercase ) _lowerCamelCase : Optional[Any] = nn.LayerNorm(config.hidden_size ) _lowerCamelCase : int = nn.Linear(config.hidden_size , self.proj_size ) # uncondition for scaling _lowerCamelCase : str = nn.Parameter(torch.randn((1, 1, self.proj_size) ) ) def A_ ( self , lowercase , lowercase=False ): _lowerCamelCase : Union[str, Any] = self.model(pixel_values=lowercase ) _lowerCamelCase : int = clip_output.pooler_output _lowerCamelCase : str = self.mapper(latent_states[:, None] ) _lowerCamelCase : List[Any] = self.final_layer_norm(lowercase ) _lowerCamelCase : Dict = self.proj_out(lowercase ) if return_uncond_vector: return latent_states, self.uncond_vector return latent_states class lowerCAmelCase__ ( nn.Module ): '''simple docstring''' def __init__( self , lowercase ): super().__init__() _lowerCamelCase : Tuple = (config.num_hidden_layers + 1) // 5 _lowerCamelCase : int = config.hidden_size _lowerCamelCase : Optional[Any] = 1 _lowerCamelCase : str = nn.ModuleList( [ BasicTransformerBlock(lowercase , lowercase , lowercase , activation_fn='gelu' , attention_bias=lowercase ) for _ in range(lowercase ) ] ) def A_ ( self , lowercase ): for block in self.blocks: _lowerCamelCase : Tuple = block(lowercase ) return hidden_states
96
1
"""simple docstring""" import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, get_image_size, is_torch_available, is_torch_tensor, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_torch_available(): import torch if is_vision_available(): import PIL lowercase__ = logging.get_logger(__name__) def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): def constraint_to_multiple_of(lowercase__ , lowercase__ , lowercase__=0 , lowercase__=None ): _lowerCamelCase : List[Any] = round(val / multiple ) * multiple if max_val is not None and x > max_val: _lowerCamelCase : Optional[int] = math.floor(val / multiple ) * multiple if x < min_val: _lowerCamelCase : List[str] = math.ceil(val / multiple ) * multiple return x _lowerCamelCase : Optional[int] = (output_size, output_size) if isinstance(lowercase__ , lowercase__ ) else output_size _lowerCamelCase, _lowerCamelCase : List[str] = get_image_size(lowercase__ ) _lowerCamelCase, _lowerCamelCase : Union[str, Any] = output_size # determine new height and width _lowerCamelCase : Dict = output_height / input_height _lowerCamelCase : Any = output_width / input_width if keep_aspect_ratio: # scale as little as possible if abs(1 - scale_width ) < abs(1 - scale_height ): # fit width _lowerCamelCase : Tuple = scale_width else: # fit height _lowerCamelCase : Any = scale_height _lowerCamelCase : Optional[int] = constraint_to_multiple_of(scale_height * input_height , multiple=lowercase__ ) _lowerCamelCase : Any = constraint_to_multiple_of(scale_width * input_width , multiple=lowercase__ ) return (new_height, new_width) class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = ["""pixel_values"""] def __init__( self , lowercase = True , lowercase = None , lowercase = PILImageResampling.BILINEAR , lowercase = False , lowercase = 1 , lowercase = True , lowercase = 1 / 255 , lowercase = True , lowercase = None , lowercase = None , **lowercase , ): super().__init__(**lowercase ) _lowerCamelCase : Optional[int] = size if size is not None else {'height': 384, 'width': 384} _lowerCamelCase : Union[str, Any] = get_size_dict(lowercase ) _lowerCamelCase : Optional[Any] = do_resize _lowerCamelCase : List[Any] = size _lowerCamelCase : Tuple = keep_aspect_ratio _lowerCamelCase : List[Any] = ensure_multiple_of _lowerCamelCase : str = resample _lowerCamelCase : Optional[Any] = do_rescale _lowerCamelCase : Union[str, Any] = rescale_factor _lowerCamelCase : int = do_normalize _lowerCamelCase : int = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _lowerCamelCase : Optional[Any] = image_std if image_std is not None else IMAGENET_STANDARD_STD def A_ ( self , lowercase , lowercase , lowercase = False , lowercase = 1 , lowercase = PILImageResampling.BICUBIC , lowercase = None , **lowercase , ): _lowerCamelCase : str = get_size_dict(lowercase ) if "height" not in size or "width" not in size: raise ValueError(F'''The size dictionary must contain the keys \'height\' and \'width\'. Got {size.keys()}''' ) _lowerCamelCase : Optional[Any] = get_resize_output_image_size( lowercase , output_size=(size['height'], size['width']) , keep_aspect_ratio=lowercase , multiple=lowercase , ) return resize(lowercase , size=lowercase , resample=lowercase , data_format=lowercase , **lowercase ) def A_ ( self , lowercase , lowercase , lowercase = None , **lowercase , ): return rescale(lowercase , scale=lowercase , data_format=lowercase , **lowercase ) def A_ ( self , lowercase , lowercase , lowercase , lowercase = None , **lowercase , ): return normalize(lowercase , mean=lowercase , std=lowercase , data_format=lowercase , **lowercase ) def A_ ( self , lowercase , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = ChannelDimension.FIRST , **lowercase , ): _lowerCamelCase : List[str] = do_resize if do_resize is not None else self.do_resize _lowerCamelCase : List[str] = size if size is not None else self.size _lowerCamelCase : Tuple = get_size_dict(lowercase ) _lowerCamelCase : Any = keep_aspect_ratio if keep_aspect_ratio is not None else self.keep_aspect_ratio _lowerCamelCase : int = ensure_multiple_of if ensure_multiple_of is not None else self.ensure_multiple_of _lowerCamelCase : int = resample if resample is not None else self.resample _lowerCamelCase : int = do_rescale if do_rescale is not None else self.do_rescale _lowerCamelCase : int = rescale_factor if rescale_factor is not None else self.rescale_factor _lowerCamelCase : Union[str, Any] = do_normalize if do_normalize is not None else self.do_normalize _lowerCamelCase : Dict = image_mean if image_mean is not None else self.image_mean _lowerCamelCase : Tuple = image_std if image_std is not None else self.image_std _lowerCamelCase : List[Any] = make_list_of_images(lowercase ) if not valid_images(lowercase ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. _lowerCamelCase : Tuple = [to_numpy_array(lowercase ) for image in images] if do_resize: _lowerCamelCase : List[Any] = [self.resize(image=lowercase , size=lowercase , resample=lowercase ) for image in images] if do_rescale: _lowerCamelCase : List[Any] = [self.rescale(image=lowercase , scale=lowercase ) for image in images] if do_normalize: _lowerCamelCase : List[str] = [self.normalize(image=lowercase , mean=lowercase , std=lowercase ) for image in images] _lowerCamelCase : Optional[Any] = [to_channel_dimension_format(lowercase , lowercase ) for image in images] _lowerCamelCase : Dict = {'pixel_values': images} return BatchFeature(data=lowercase , tensor_type=lowercase ) def A_ ( self , lowercase , lowercase = None ): _lowerCamelCase : int = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(lowercase ) != len(lowercase ): raise ValueError( 'Make sure that you pass in as many target sizes as the batch dimension of the logits' ) if is_torch_tensor(lowercase ): _lowerCamelCase : Union[str, Any] = target_sizes.numpy() _lowerCamelCase : Union[str, Any] = [] for idx in range(len(lowercase ) ): _lowerCamelCase : Tuple = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode='bilinear' , align_corners=lowercase ) _lowerCamelCase : Dict = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(lowercase ) else: _lowerCamelCase : Dict = logits.argmax(dim=1 ) _lowerCamelCase : int = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
96
"""simple docstring""" lowercase__ = { """meter""": """m""", """kilometer""": """km""", """megametre""": """Mm""", """gigametre""": """Gm""", """terametre""": """Tm""", """petametre""": """Pm""", """exametre""": """Em""", """zettametre""": """Zm""", """yottametre""": """Ym""", } # Exponent of the factor(meter) lowercase__ = { """m""": 0, """km""": 3, """Mm""": 6, """Gm""": 9, """Tm""": 12, """Pm""": 15, """Em""": 18, """Zm""": 21, """Ym""": 24, } def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : List[Any] = from_type.lower().strip('s' ) _lowerCamelCase : List[Any] = to_type.lower().strip('s' ) _lowerCamelCase : Optional[int] = UNIT_SYMBOL.get(lowercase__ , lowercase__ ) _lowerCamelCase : Any = UNIT_SYMBOL.get(lowercase__ , lowercase__ ) if from_sanitized not in METRIC_CONVERSION: _lowerCamelCase : Tuple = ( f'''Invalid \'from_type\' value: {from_type!r}.\n''' f'''Conversion abbreviations are: {', '.join(lowercase__ )}''' ) raise ValueError(lowercase__ ) if to_sanitized not in METRIC_CONVERSION: _lowerCamelCase : Any = ( f'''Invalid \'to_type\' value: {to_type!r}.\n''' f'''Conversion abbreviations are: {', '.join(lowercase__ )}''' ) raise ValueError(lowercase__ ) _lowerCamelCase : List[Any] = METRIC_CONVERSION[from_sanitized] _lowerCamelCase : int = METRIC_CONVERSION[to_sanitized] _lowerCamelCase : List[str] = 1 if from_exponent > to_exponent: _lowerCamelCase : List[str] = from_exponent - to_exponent else: _lowerCamelCase : List[Any] = -(to_exponent - from_exponent) return value * pow(10 , lowercase__ ) if __name__ == "__main__": from doctest import testmod testmod()
96
1
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator from typing import Any class lowerCAmelCase__ : '''simple docstring''' def __init__( self , lowercase ): _lowerCamelCase : Any = data _lowerCamelCase : Node | None = None class lowerCAmelCase__ : '''simple docstring''' def __init__( self ): _lowerCamelCase : str = None _lowerCamelCase : str = None def __iter__( self ): _lowerCamelCase : List[str] = self.head while self.head: yield node.data _lowerCamelCase : Optional[int] = node.next if node == self.head: break def __len__( self ): return sum(1 for _ in self ) def __repr__( self ): return "->".join(str(lowercase ) for item in iter(self ) ) def A_ ( self , lowercase ): self.insert_nth(len(self ) , lowercase ) def A_ ( self , lowercase ): self.insert_nth(0 , lowercase ) def A_ ( self , lowercase , lowercase ): if index < 0 or index > len(self ): raise IndexError('list index out of range.' ) _lowerCamelCase : List[Any] = Node(lowercase ) if self.head is None: _lowerCamelCase : str = new_node # first node points itself _lowerCamelCase : Union[str, Any] = new_node elif index == 0: # insert at head _lowerCamelCase : List[str] = self.head _lowerCamelCase : str = new_node else: _lowerCamelCase : Union[str, Any] = self.head for _ in range(index - 1 ): _lowerCamelCase : List[Any] = temp.next _lowerCamelCase : Union[str, Any] = temp.next _lowerCamelCase : List[str] = new_node if index == len(self ) - 1: # insert at tail _lowerCamelCase : Any = new_node def A_ ( self ): return self.delete_nth(0 ) def A_ ( self ): return self.delete_nth(len(self ) - 1 ) def A_ ( self , lowercase = 0 ): if not 0 <= index < len(self ): raise IndexError('list index out of range.' ) _lowerCamelCase : Any = self.head if self.head == self.tail: # just one node _lowerCamelCase : List[str] = None elif index == 0: # delete head node _lowerCamelCase : List[str] = self.tail.next.next _lowerCamelCase : Optional[int] = self.head.next else: _lowerCamelCase : Dict = self.head for _ in range(index - 1 ): _lowerCamelCase : List[Any] = temp.next _lowerCamelCase : int = temp.next _lowerCamelCase : Optional[int] = temp.next.next if index == len(self ) - 1: # delete at tail _lowerCamelCase : List[Any] = temp return delete_node.data def A_ ( self ): return len(self ) == 0 def _snake_case ( ): _lowerCamelCase : Union[str, Any] = CircularLinkedList() assert len(lowercase__ ) == 0 assert circular_linked_list.is_empty() is True assert str(lowercase__ ) == "" try: circular_linked_list.delete_front() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_tail() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_nth(-1 ) raise AssertionError except IndexError: assert True try: circular_linked_list.delete_nth(0 ) raise AssertionError except IndexError: assert True assert circular_linked_list.is_empty() is True for i in range(5 ): assert len(lowercase__ ) == i circular_linked_list.insert_nth(lowercase__ , i + 1 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 6 ) ) circular_linked_list.insert_tail(6 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 7 ) ) circular_linked_list.insert_head(0 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(0 , 7 ) ) assert circular_linked_list.delete_front() == 0 assert circular_linked_list.delete_tail() == 6 assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 6 ) ) assert circular_linked_list.delete_nth(2 ) == 3 circular_linked_list.insert_nth(2 , 3 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 6 ) ) assert circular_linked_list.is_empty() is False if __name__ == "__main__": import doctest doctest.testmod()
96
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ = {"""configuration_vit_msn""": ["""VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ViTMSNConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ """VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST""", """ViTMSNModel""", """ViTMSNForImageClassification""", """ViTMSNPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_vit_msn import VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMSNConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_msn import ( VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMSNForImageClassification, ViTMSNModel, ViTMSNPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
96
1
"""simple docstring""" import argparse import json from collections import OrderedDict from functools import partial from pathlib import Path import timm import torch from huggingface_hub import hf_hub_download from transformers import LevitConfig, LevitForImageClassificationWithTeacher, LevitImageProcessor from transformers.utils import logging logging.set_verbosity_info() lowercase__ = logging.get_logger() def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ = True ): print(f'''Converting {name}...''' ) with torch.no_grad(): if hidden_sizes == 128: if name[-1] == "S": _lowerCamelCase : Dict = timm.create_model('levit_128s' , pretrained=lowercase__ ) else: _lowerCamelCase : int = timm.create_model('levit_128' , pretrained=lowercase__ ) if hidden_sizes == 192: _lowerCamelCase : Optional[int] = timm.create_model('levit_192' , pretrained=lowercase__ ) if hidden_sizes == 256: _lowerCamelCase : Optional[Any] = timm.create_model('levit_256' , pretrained=lowercase__ ) if hidden_sizes == 384: _lowerCamelCase : str = timm.create_model('levit_384' , pretrained=lowercase__ ) from_model.eval() _lowerCamelCase : List[Any] = LevitForImageClassificationWithTeacher(lowercase__ ).eval() _lowerCamelCase : Optional[int] = OrderedDict() _lowerCamelCase : List[Any] = from_model.state_dict() _lowerCamelCase : List[Any] = list(from_model.state_dict().keys() ) _lowerCamelCase : Union[str, Any] = list(our_model.state_dict().keys() ) print(len(lowercase__ ) , len(lowercase__ ) ) for i in range(len(lowercase__ ) ): _lowerCamelCase : Union[str, Any] = weights[og_keys[i]] our_model.load_state_dict(lowercase__ ) _lowerCamelCase : Optional[Any] = torch.randn((2, 3, 224, 224) ) _lowerCamelCase : Union[str, Any] = from_model(lowercase__ ) _lowerCamelCase : Tuple = our_model(lowercase__ ).logits assert torch.allclose(lowercase__ , lowercase__ ), "The model logits don't match the original one." _lowerCamelCase : Any = name print(lowercase__ ) if push_to_hub: our_model.save_pretrained(save_directory / checkpoint_name ) _lowerCamelCase : List[Any] = LevitImageProcessor() image_processor.save_pretrained(save_directory / checkpoint_name ) print(f'''Pushed {checkpoint_name}''' ) def _snake_case ( lowercase__ , lowercase__ = None , lowercase__ = True ): _lowerCamelCase : str = 'imagenet-1k-id2label.json' _lowerCamelCase : Dict = 1000 _lowerCamelCase : Union[str, Any] = (1, num_labels) _lowerCamelCase : Optional[Any] = 'huggingface/label-files' _lowerCamelCase : Tuple = num_labels _lowerCamelCase : str = json.load(open(hf_hub_download(lowercase__ , lowercase__ , repo_type='dataset' ) , 'r' ) ) _lowerCamelCase : List[Any] = {int(lowercase__ ): v for k, v in idalabel.items()} _lowerCamelCase : int = idalabel _lowerCamelCase : Dict = {v: k for k, v in idalabel.items()} _lowerCamelCase : int = partial(lowercase__ , num_labels=lowercase__ , idalabel=lowercase__ , labelaid=lowercase__ ) _lowerCamelCase : Optional[Any] = { 'levit-128S': 128, 'levit-128': 128, 'levit-192': 192, 'levit-256': 256, 'levit-384': 384, } _lowerCamelCase : Optional[Any] = { 'levit-128S': ImageNetPreTrainedConfig( hidden_sizes=[128, 256, 384] , num_attention_heads=[4, 6, 8] , depths=[2, 3, 4] , key_dim=[16, 16, 16] , drop_path_rate=0 , ), 'levit-128': ImageNetPreTrainedConfig( hidden_sizes=[128, 256, 384] , num_attention_heads=[4, 8, 12] , depths=[4, 4, 4] , key_dim=[16, 16, 16] , drop_path_rate=0 , ), 'levit-192': ImageNetPreTrainedConfig( hidden_sizes=[192, 288, 384] , num_attention_heads=[3, 5, 6] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0 , ), 'levit-256': ImageNetPreTrainedConfig( hidden_sizes=[256, 384, 512] , num_attention_heads=[4, 6, 8] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0 , ), 'levit-384': ImageNetPreTrainedConfig( hidden_sizes=[384, 512, 768] , num_attention_heads=[6, 9, 12] , depths=[4, 4, 4] , key_dim=[32, 32, 32] , drop_path_rate=0.1 , ), } if model_name: convert_weight_and_push( names_to_hidden_sizes[model_name] , lowercase__ , names_to_config[model_name] , lowercase__ , lowercase__ ) else: for model_name, config in names_to_config.items(): convert_weight_and_push(names_to_hidden_sizes[model_name] , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) return config, expected_shape if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default=None, type=str, help="""The name of the model you wish to convert, it must be one of the supported Levit* architecture,""", ) parser.add_argument( """--pytorch_dump_folder_path""", default="""levit-dump-folder/""", type=Path, required=False, help="""Path to the output PyTorch model directory.""", ) parser.add_argument("""--push_to_hub""", action="""store_true""", help="""Push model and image processor to the hub""") parser.add_argument( """--no-push_to_hub""", dest="""push_to_hub""", action="""store_false""", help="""Do not push model and image processor to the hub""", ) lowercase__ = parser.parse_args() lowercase__ = args.pytorch_dump_folder_path pytorch_dump_folder_path.mkdir(exist_ok=True, parents=True) convert_weights_and_push(pytorch_dump_folder_path, args.model_name, args.push_to_hub)
96
"""simple docstring""" def _snake_case ( lowercase__ , lowercase__ ): # "extended trapezoidal rule" # int(f) = dx/2 * (f1 + 2f2 + ... + fn) _lowerCamelCase : List[Any] = (boundary[1] - boundary[0]) / steps _lowerCamelCase : Tuple = boundary[0] _lowerCamelCase : Dict = boundary[1] _lowerCamelCase : List[Any] = make_points(lowercase__ , lowercase__ , lowercase__ ) _lowerCamelCase : List[Any] = 0.0 y += (h / 2.0) * f(lowercase__ ) for i in x_i: # print(i) y += h * f(lowercase__ ) y += (h / 2.0) * f(lowercase__ ) return y def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : str = a + h while x < (b - h): yield x _lowerCamelCase : int = x + h def _snake_case ( lowercase__ ): # enter your function here _lowerCamelCase : Optional[Any] = (x - 0) * (x - 0) return y def _snake_case ( ): _lowerCamelCase : int = 0.0 # Lower bound of integration _lowerCamelCase : Optional[int] = 1.0 # Upper bound of integration _lowerCamelCase : List[str] = 1_0.0 # define number of steps or resolution _lowerCamelCase : List[Any] = [a, b] # define boundary of integration _lowerCamelCase : Optional[Any] = method_a(lowercase__ , lowercase__ ) print(f'''y = {y}''' ) if __name__ == "__main__": main()
96
1
"""simple docstring""" import math import unittest from transformers import BioGptConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptTokenizer, ) from transformers.models.biogpt.modeling_biogpt import BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCAmelCase__ : '''simple docstring''' def __init__( self , lowercase , lowercase=13 , lowercase=7 , lowercase=True , lowercase=True , lowercase=False , lowercase=True , lowercase=99 , lowercase=32 , lowercase=5 , lowercase=4 , lowercase=37 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=512 , lowercase=16 , lowercase=2 , lowercase=0.02 , lowercase=3 , lowercase=4 , lowercase=None , ): _lowerCamelCase : List[Any] = parent _lowerCamelCase : Optional[int] = batch_size _lowerCamelCase : Optional[Any] = seq_length _lowerCamelCase : Optional[int] = is_training _lowerCamelCase : Tuple = use_input_mask _lowerCamelCase : str = use_token_type_ids _lowerCamelCase : List[Any] = use_labels _lowerCamelCase : int = vocab_size _lowerCamelCase : str = hidden_size _lowerCamelCase : Tuple = num_hidden_layers _lowerCamelCase : Any = num_attention_heads _lowerCamelCase : Any = intermediate_size _lowerCamelCase : Dict = hidden_act _lowerCamelCase : str = hidden_dropout_prob _lowerCamelCase : Optional[int] = attention_probs_dropout_prob _lowerCamelCase : List[Any] = max_position_embeddings _lowerCamelCase : int = type_vocab_size _lowerCamelCase : Dict = type_sequence_label_size _lowerCamelCase : Union[str, Any] = initializer_range _lowerCamelCase : int = num_labels _lowerCamelCase : Tuple = num_choices _lowerCamelCase : int = scope def A_ ( self ): _lowerCamelCase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase : Tuple = None if self.use_input_mask: _lowerCamelCase : str = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCamelCase : Any = None if self.use_token_type_ids: _lowerCamelCase : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCamelCase : Dict = None _lowerCamelCase : Optional[Any] = None _lowerCamelCase : Dict = None if self.use_labels: _lowerCamelCase : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase : Any = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _lowerCamelCase : Dict = ids_tensor([self.batch_size] , self.num_choices ) _lowerCamelCase : List[str] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A_ ( self ): return BioGptConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowercase , initializer_range=self.initializer_range , ) def A_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ): _lowerCamelCase : Optional[Any] = BioGptModel(config=lowercase ) model.to(lowercase ) model.eval() _lowerCamelCase : List[str] = model(lowercase , attention_mask=lowercase ) _lowerCamelCase : Optional[int] = model(lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , ): _lowerCamelCase : int = BioGptForCausalLM(config=lowercase ) model.to(lowercase ) model.eval() _lowerCamelCase : Optional[Any] = model(lowercase , attention_mask=lowercase , token_type_ids=lowercase , labels=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase ): _lowerCamelCase : List[str] = BioGptModel(config=lowercase ) model.to(lowercase ) model.eval() # create attention mask _lowerCamelCase : List[str] = torch.ones(input_ids.shape , dtype=torch.long , device=lowercase ) _lowerCamelCase : Optional[int] = self.seq_length // 2 _lowerCamelCase : List[str] = 0 # first forward pass _lowerCamelCase, _lowerCamelCase : Optional[int] = model(lowercase , attention_mask=lowercase ).to_tuple() # create hypothetical next token and extent to next_input_ids _lowerCamelCase : Union[str, Any] = ids_tensor((self.batch_size, 1) , config.vocab_size ) # change a random masked slice from input_ids _lowerCamelCase : Tuple = ids_tensor((1,) , lowercase ).item() + 1 _lowerCamelCase : Tuple = ids_tensor((self.batch_size, 1) , config.vocab_size ).squeeze(-1 ) _lowerCamelCase : str = random_other_next_tokens # append to next input_ids and attn_mask _lowerCamelCase : List[str] = torch.cat([input_ids, next_tokens] , dim=-1 ) _lowerCamelCase : Optional[int] = torch.cat( [attn_mask, torch.ones((attn_mask.shape[0], 1) , dtype=torch.long , device=lowercase )] , dim=1 , ) # get two different outputs _lowerCamelCase : str = model(lowercase , attention_mask=lowercase )['last_hidden_state'] _lowerCamelCase : str = model(lowercase , past_key_values=lowercase , attention_mask=lowercase )['last_hidden_state'] # select random slice _lowerCamelCase : Dict = ids_tensor((1,) , output_from_past.shape[-1] ).item() _lowerCamelCase : Union[str, Any] = output_from_no_past[:, -1, random_slice_idx].detach() _lowerCamelCase : Tuple = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowercase , lowercase , atol=1E-3 ) ) def A_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase ): _lowerCamelCase : int = BioGptModel(config=lowercase ).to(lowercase ).eval() _lowerCamelCase : List[str] = torch.ones(input_ids.shape , dtype=torch.long , device=lowercase ) # first forward pass _lowerCamelCase : List[Any] = model(lowercase , attention_mask=lowercase , use_cache=lowercase ) _lowerCamelCase, _lowerCamelCase : Dict = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids _lowerCamelCase : Optional[Any] = ids_tensor((self.batch_size, 3) , config.vocab_size ) _lowerCamelCase : Optional[int] = ids_tensor((self.batch_size, 3) , 2 ) # append to next input_ids and _lowerCamelCase : Tuple = torch.cat([input_ids, next_tokens] , dim=-1 ) _lowerCamelCase : Optional[Any] = torch.cat([attention_mask, next_attn_mask] , dim=-1 ) _lowerCamelCase : List[Any] = model(lowercase , attention_mask=lowercase )['last_hidden_state'] _lowerCamelCase : Dict = model(lowercase , attention_mask=lowercase , past_key_values=lowercase )[ 'last_hidden_state' ] # select random slice _lowerCamelCase : Optional[Any] = ids_tensor((1,) , output_from_past.shape[-1] ).item() _lowerCamelCase : List[str] = output_from_no_past[:, -3:, random_slice_idx].detach() _lowerCamelCase : int = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowercase , lowercase , atol=1E-3 ) ) def A_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase , lowercase=False ): _lowerCamelCase : List[str] = BioGptForCausalLM(lowercase ) model.to(lowercase ) if gradient_checkpointing: model.gradient_checkpointing_enable() _lowerCamelCase : List[Any] = model(lowercase , labels=lowercase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) result.loss.backward() def A_ ( self , lowercase , *lowercase ): _lowerCamelCase : List[str] = BioGptModel(lowercase ) _lowerCamelCase : Optional[Any] = model.config.initializer_range / math.sqrt(2 * model.config.num_hidden_layers ) for key in model.state_dict().keys(): if "c_proj" in key and "weight" in key: self.parent.assertLessEqual(abs(torch.std(model.state_dict()[key] ) - model_std ) , 0.0_01 ) self.parent.assertLessEqual(abs(torch.mean(model.state_dict()[key] ) - 0.0 ) , 0.01 ) def A_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase ): _lowerCamelCase : Dict = self.num_labels _lowerCamelCase : int = BioGptForTokenClassification(lowercase ) model.to(lowercase ) model.eval() _lowerCamelCase : List[Any] = model(lowercase , attention_mask=lowercase , token_type_ids=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A_ ( self ): _lowerCamelCase : str = self.prepare_config_and_inputs() ( ( _lowerCamelCase ), ( _lowerCamelCase ), ( _lowerCamelCase ), ( _lowerCamelCase ), ( _lowerCamelCase ), ( _lowerCamelCase ), ( _lowerCamelCase ), ) : Dict = config_and_inputs _lowerCamelCase : str = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class lowerCAmelCase__ ( lowercase, lowercase, lowercase, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = ( (BioGptModel, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification) if is_torch_available() else () ) lowerCamelCase__ = (BioGptForCausalLM,) if is_torch_available() else () lowerCamelCase__ = ( { """feature-extraction""": BioGptModel, """text-classification""": BioGptForSequenceClassification, """text-generation""": BioGptForCausalLM, """token-classification""": BioGptForTokenClassification, """zero-shot""": BioGptForSequenceClassification, } if is_torch_available() else {} ) lowerCamelCase__ = False def A_ ( self ): _lowerCamelCase : Tuple = BioGptModelTester(self ) _lowerCamelCase : Optional[Any] = ConfigTester(self , config_class=lowercase , hidden_size=37 ) def A_ ( self ): self.config_tester.run_common_tests() def A_ ( self ): _lowerCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase ) def A_ ( self ): _lowerCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _lowerCamelCase : Tuple = type self.model_tester.create_and_check_model(*lowercase ) def A_ ( self ): _lowerCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_attention_mask_past(*lowercase ) def A_ ( self ): _lowerCamelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_forward_and_backwards(*lowercase , gradient_checkpointing=lowercase ) def A_ ( self ): _lowerCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_past_large_inputs(*lowercase ) def A_ ( self ): _lowerCamelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_weight_initialization(*lowercase ) def A_ ( self ): _lowerCamelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_for_token_classification(*lowercase ) @slow def A_ ( self ): _lowerCamelCase : Optional[int] = BioGptForCausalLM.from_pretrained('microsoft/biogpt' ) model.to(lowercase ) _lowerCamelCase : List[str] = BioGptTokenizer.from_pretrained('microsoft/biogpt' ) _lowerCamelCase : Optional[int] = 'left' # Define PAD Token = EOS Token = 50256 _lowerCamelCase : Tuple = tokenizer.eos_token _lowerCamelCase : Optional[int] = model.config.eos_token_id # use different length sentences to test batching _lowerCamelCase : Union[str, Any] = [ 'Hello, my dog is a little', 'Today, I', ] _lowerCamelCase : Optional[Any] = tokenizer(lowercase , return_tensors='pt' , padding=lowercase ) _lowerCamelCase : Dict = inputs['input_ids'].to(lowercase ) _lowerCamelCase : Union[str, Any] = model.generate( input_ids=lowercase , attention_mask=inputs['attention_mask'].to(lowercase ) , ) _lowerCamelCase : Union[str, Any] = tokenizer(sentences[0] , return_tensors='pt' ).input_ids.to(lowercase ) _lowerCamelCase : List[Any] = model.generate(input_ids=lowercase ) _lowerCamelCase : List[str] = inputs_non_padded.shape[-1] - inputs['attention_mask'][-1].long().sum().cpu().item() _lowerCamelCase : Any = tokenizer(sentences[1] , return_tensors='pt' ).input_ids.to(lowercase ) _lowerCamelCase : List[str] = model.generate(input_ids=lowercase , max_length=model.config.max_length - num_paddings ) _lowerCamelCase : int = tokenizer.batch_decode(lowercase , skip_special_tokens=lowercase ) _lowerCamelCase : List[Any] = tokenizer.decode(output_non_padded[0] , skip_special_tokens=lowercase ) _lowerCamelCase : int = tokenizer.decode(output_padded[0] , skip_special_tokens=lowercase ) _lowerCamelCase : str = [ 'Hello, my dog is a little bit bigger than a little bit.', 'Today, I have a good idea of how to use the information', ] self.assertListEqual(lowercase , lowercase ) self.assertListEqual(lowercase , [non_padded_sentence, padded_sentence] ) @slow def A_ ( self ): for model_name in BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : List[Any] = BioGptModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) def A_ ( self ): _lowerCamelCase, _lowerCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase : Union[str, Any] = 3 _lowerCamelCase : Optional[int] = input_dict['input_ids'] _lowerCamelCase : Tuple = input_ids.ne(1 ).to(lowercase ) _lowerCamelCase : Any = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) _lowerCamelCase : Tuple = BioGptForSequenceClassification(lowercase ) model.to(lowercase ) model.eval() _lowerCamelCase : str = model(lowercase , attention_mask=lowercase , labels=lowercase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A_ ( self ): _lowerCamelCase, _lowerCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase : int = 3 _lowerCamelCase : List[Any] = 'multi_label_classification' _lowerCamelCase : Tuple = input_dict['input_ids'] _lowerCamelCase : List[Any] = input_ids.ne(1 ).to(lowercase ) _lowerCamelCase : str = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) _lowerCamelCase : Union[str, Any] = BioGptForSequenceClassification(lowercase ) model.to(lowercase ) model.eval() _lowerCamelCase : List[str] = model(lowercase , attention_mask=lowercase , labels=lowercase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @require_torch class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' @slow def A_ ( self ): _lowerCamelCase : str = BioGptForCausalLM.from_pretrained('microsoft/biogpt' ) _lowerCamelCase : List[str] = torch.tensor([[2, 4805, 9, 656, 21]] ) _lowerCamelCase : List[str] = model(lowercase )[0] _lowerCamelCase : str = 42384 _lowerCamelCase : int = torch.Size((1, 5, vocab_size) ) self.assertEqual(output.shape , lowercase ) _lowerCamelCase : str = torch.tensor( [[[-9.52_36, -9.89_18, 10.45_57], [-11.04_69, -9.64_23, 8.10_22], [-8.86_64, -7.88_26, 5.53_25]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , lowercase , atol=1E-4 ) ) @slow def A_ ( self ): _lowerCamelCase : Union[str, Any] = BioGptTokenizer.from_pretrained('microsoft/biogpt' ) _lowerCamelCase : Tuple = BioGptForCausalLM.from_pretrained('microsoft/biogpt' ) model.to(lowercase ) torch.manual_seed(0 ) _lowerCamelCase : Union[str, Any] = tokenizer('COVID-19 is' , return_tensors='pt' ).to(lowercase ) _lowerCamelCase : int = model.generate( **lowercase , min_length=100 , max_length=1024 , num_beams=5 , early_stopping=lowercase , ) _lowerCamelCase : Any = tokenizer.decode(output_ids[0] , skip_special_tokens=lowercase ) _lowerCamelCase : List[str] = ( 'COVID-19 is a global pandemic caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), the' ' causative agent of coronavirus disease 2019 (COVID-19), which has spread to more than 200 countries and' ' territories, including the United States (US), Canada, Australia, New Zealand, the United Kingdom (UK),' ' and the United States of America (USA), as of March 11, 2020, with more than 800,000 confirmed cases and' ' more than 800,000 deaths.' ) self.assertEqual(lowercase , lowercase )
96
"""simple docstring""" import math def _snake_case ( lowercase__ ): return math.sqrt(lowercase__ ) * math.sqrt(lowercase__ ) == num def _snake_case ( lowercase__ ): _lowerCamelCase : Optional[int] = 0 _lowerCamelCase : List[Any] = n while left <= right: _lowerCamelCase : str = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: _lowerCamelCase : str = mid - 1 else: _lowerCamelCase : Optional[int] = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" import argparse import os import torch from transformers import ( XLNetConfig, XLNetForQuestionAnswering, XLNetForSequenceClassification, XLNetLMHeadModel, load_tf_weights_in_xlnet, ) from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging lowercase__ = { """cola""": 2, """mnli""": 3, """mrpc""": 2, """sst-2""": 2, """sts-b""": 1, """qqp""": 2, """qnli""": 2, """rte""": 2, """wnli""": 2, } logging.set_verbosity_info() def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__=None ): # Initialise PyTorch model _lowerCamelCase : Dict = XLNetConfig.from_json_file(lowercase__ ) _lowerCamelCase : int = finetuning_task.lower() if finetuning_task is not None else '' if finetuning_task in GLUE_TASKS_NUM_LABELS: print(f'''Building PyTorch XLNetForSequenceClassification model from configuration: {config}''' ) _lowerCamelCase : int = finetuning_task _lowerCamelCase : str = GLUE_TASKS_NUM_LABELS[finetuning_task] _lowerCamelCase : Dict = XLNetForSequenceClassification(lowercase__ ) elif "squad" in finetuning_task: _lowerCamelCase : Any = finetuning_task _lowerCamelCase : Dict = XLNetForQuestionAnswering(lowercase__ ) else: _lowerCamelCase : Tuple = XLNetLMHeadModel(lowercase__ ) # Load weights from tf checkpoint load_tf_weights_in_xlnet(lowercase__ , lowercase__ , lowercase__ ) # Save pytorch-model _lowerCamelCase : Tuple = os.path.join(lowercase__ , lowercase__ ) _lowerCamelCase : Any = os.path.join(lowercase__ , lowercase__ ) print(f'''Save PyTorch model to {os.path.abspath(lowercase__ )}''' ) torch.save(model.state_dict() , lowercase__ ) print(f'''Save configuration file to {os.path.abspath(lowercase__ )}''' ) with open(lowercase__ , 'w' , encoding='utf-8' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--xlnet_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained XLNet model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the folder to store the PyTorch model or dataset/vocab.""", ) parser.add_argument( """--finetuning_task""", default=None, type=str, help="""Name of a task on which the XLNet TensorFlow model was fine-tuned""", ) lowercase__ = parser.parse_args() print(args) convert_xlnet_checkpoint_to_pytorch( args.tf_checkpoint_path, args.xlnet_config_file, args.pytorch_dump_folder_path, args.finetuning_task )
96
"""simple docstring""" import functools from typing import Any def _snake_case ( lowercase__ , lowercase__ ): # Validation if not isinstance(lowercase__ , lowercase__ ) or len(lowercase__ ) == 0: raise ValueError('the string should be not empty string' ) if not isinstance(lowercase__ , lowercase__ ) or not all( isinstance(lowercase__ , lowercase__ ) and len(lowercase__ ) > 0 for item in words ): raise ValueError('the words should be a list of non-empty strings' ) # Build trie _lowerCamelCase : dict[str, Any] = {} _lowerCamelCase : List[Any] = 'WORD_KEEPER' for word in words: _lowerCamelCase : Dict = trie for c in word: if c not in trie_node: _lowerCamelCase : Any = {} _lowerCamelCase : str = trie_node[c] _lowerCamelCase : Optional[Any] = True _lowerCamelCase : Dict = len(lowercase__ ) # Dynamic programming method @functools.cache def is_breakable(lowercase__ ) -> bool: if index == len_string: return True _lowerCamelCase : List[Any] = trie for i in range(lowercase__ , lowercase__ ): _lowerCamelCase : Any = trie_node.get(string[i] , lowercase__ ) if trie_node is None: return False if trie_node.get(lowercase__ , lowercase__ ) and is_breakable(i + 1 ): return True return False return is_breakable(0 ) if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_funnel import FunnelTokenizer lowercase__ = logging.get_logger(__name__) lowercase__ = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} lowercase__ = [ """small""", """small-base""", """medium""", """medium-base""", """intermediate""", """intermediate-base""", """large""", """large-base""", """xlarge""", """xlarge-base""", ] lowercase__ = { """vocab_file""": { """funnel-transformer/small""": """https://huggingface.co/funnel-transformer/small/resolve/main/vocab.txt""", """funnel-transformer/small-base""": """https://huggingface.co/funnel-transformer/small-base/resolve/main/vocab.txt""", """funnel-transformer/medium""": """https://huggingface.co/funnel-transformer/medium/resolve/main/vocab.txt""", """funnel-transformer/medium-base""": ( """https://huggingface.co/funnel-transformer/medium-base/resolve/main/vocab.txt""" ), """funnel-transformer/intermediate""": ( """https://huggingface.co/funnel-transformer/intermediate/resolve/main/vocab.txt""" ), """funnel-transformer/intermediate-base""": ( """https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/vocab.txt""" ), """funnel-transformer/large""": """https://huggingface.co/funnel-transformer/large/resolve/main/vocab.txt""", """funnel-transformer/large-base""": """https://huggingface.co/funnel-transformer/large-base/resolve/main/vocab.txt""", """funnel-transformer/xlarge""": """https://huggingface.co/funnel-transformer/xlarge/resolve/main/vocab.txt""", """funnel-transformer/xlarge-base""": ( """https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """funnel-transformer/small""": """https://huggingface.co/funnel-transformer/small/resolve/main/tokenizer.json""", """funnel-transformer/small-base""": ( """https://huggingface.co/funnel-transformer/small-base/resolve/main/tokenizer.json""" ), """funnel-transformer/medium""": """https://huggingface.co/funnel-transformer/medium/resolve/main/tokenizer.json""", """funnel-transformer/medium-base""": ( """https://huggingface.co/funnel-transformer/medium-base/resolve/main/tokenizer.json""" ), """funnel-transformer/intermediate""": ( """https://huggingface.co/funnel-transformer/intermediate/resolve/main/tokenizer.json""" ), """funnel-transformer/intermediate-base""": ( """https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/tokenizer.json""" ), """funnel-transformer/large""": """https://huggingface.co/funnel-transformer/large/resolve/main/tokenizer.json""", """funnel-transformer/large-base""": ( """https://huggingface.co/funnel-transformer/large-base/resolve/main/tokenizer.json""" ), """funnel-transformer/xlarge""": """https://huggingface.co/funnel-transformer/xlarge/resolve/main/tokenizer.json""", """funnel-transformer/xlarge-base""": ( """https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/tokenizer.json""" ), }, } lowercase__ = {F"funnel-transformer/{name}": 512 for name in _model_names} lowercase__ = {F"funnel-transformer/{name}": {"""do_lower_case""": True} for name in _model_names} class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = VOCAB_FILES_NAMES lowerCamelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ = PRETRAINED_INIT_CONFIGURATION lowerCamelCase__ = FunnelTokenizer lowerCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ = 2 def __init__( self , lowercase=None , lowercase=None , lowercase=True , lowercase="<unk>" , lowercase="<sep>" , lowercase="<pad>" , lowercase="<cls>" , lowercase="<mask>" , lowercase="<s>" , lowercase="</s>" , lowercase=True , lowercase=True , lowercase=None , lowercase="##" , **lowercase , ): super().__init__( lowercase , tokenizer_file=lowercase , do_lower_case=lowercase , unk_token=lowercase , sep_token=lowercase , pad_token=lowercase , cls_token=lowercase , mask_token=lowercase , bos_token=lowercase , eos_token=lowercase , clean_text=lowercase , tokenize_chinese_chars=lowercase , strip_accents=lowercase , wordpieces_prefix=lowercase , **lowercase , ) _lowerCamelCase : Any = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('lowercase' , lowercase ) != do_lower_case or normalizer_state.get('strip_accents' , lowercase ) != strip_accents or normalizer_state.get('handle_chinese_chars' , lowercase ) != tokenize_chinese_chars ): _lowerCamelCase : List[Any] = getattr(lowercase , normalizer_state.pop('type' ) ) _lowerCamelCase : List[Any] = do_lower_case _lowerCamelCase : Any = strip_accents _lowerCamelCase : Dict = tokenize_chinese_chars _lowerCamelCase : Any = normalizer_class(**lowercase ) _lowerCamelCase : Optional[Any] = do_lower_case def A_ ( self , lowercase , lowercase=None ): _lowerCamelCase : str = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A_ ( self , lowercase , lowercase = None ): _lowerCamelCase : Any = [self.sep_token_id] _lowerCamelCase : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A_ ( self , lowercase , lowercase = None ): _lowerCamelCase : Tuple = self._tokenizer.model.save(lowercase , name=lowercase ) return tuple(lowercase )
96
"""simple docstring""" def _snake_case ( lowercase__ ): if not isinstance(lowercase__ , lowercase__ ): raise ValueError('Input series is not valid, valid series - [2, 4, 6]' ) if len(lowercase__ ) == 0: raise ValueError('Input list must be a non empty list' ) if len(lowercase__ ) == 1: return True _lowerCamelCase : List[Any] = series[1] - series[0] for index in range(len(lowercase__ ) - 1 ): if series[index + 1] - series[index] != common_diff: return False return True def _snake_case ( lowercase__ ): if not isinstance(lowercase__ , lowercase__ ): raise ValueError('Input series is not valid, valid series - [2, 4, 6]' ) if len(lowercase__ ) == 0: raise ValueError('Input list must be a non empty list' ) _lowerCamelCase : Optional[int] = 0 for val in series: answer += val return answer / len(lowercase__ ) if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_dpt import DPTImageProcessor lowercase__ = logging.get_logger(__name__) class lowerCAmelCase__ ( lowercase ): '''simple docstring''' def __init__( self , *lowercase , **lowercase ): warnings.warn( 'The class DPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use DPTImageProcessor instead.' , lowercase , ) super().__init__(*lowercase , **lowercase )
96
"""simple docstring""" import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler lowercase__ = 16 lowercase__ = 32 def _snake_case ( lowercase__ , lowercase__ = 16 , lowercase__ = "bert-base-cased" ): _lowerCamelCase : List[Any] = AutoTokenizer.from_pretrained(lowercase__ ) _lowerCamelCase : Tuple = load_dataset('glue' , 'mrpc' ) def tokenize_function(lowercase__ ): # max_length=None => use the model max length (it's actually the default) _lowerCamelCase : Union[str, Any] = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=lowercase__ , max_length=lowercase__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset _lowerCamelCase : int = datasets.map( lowercase__ , batched=lowercase__ , remove_columns=['idx', 'sentence1', 'sentence2'] , load_from_cache_file=lowercase__ ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _lowerCamelCase : Optional[int] = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(lowercase__ ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(lowercase__ , padding='max_length' , max_length=128 , return_tensors='pt' ) return tokenizer.pad(lowercase__ , padding='longest' , return_tensors='pt' ) # Instantiate dataloaders. _lowerCamelCase : List[str] = DataLoader( tokenized_datasets['train'] , shuffle=lowercase__ , collate_fn=lowercase__ , batch_size=lowercase__ ) _lowerCamelCase : int = DataLoader( tokenized_datasets['validation'] , shuffle=lowercase__ , collate_fn=lowercase__ , batch_size=lowercase__ ) return train_dataloader, eval_dataloader def _snake_case ( lowercase__ , lowercase__ ): # Initialize accelerator _lowerCamelCase : Optional[int] = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _lowerCamelCase : Optional[int] = config['lr'] _lowerCamelCase : Optional[int] = int(config['num_epochs'] ) _lowerCamelCase : Union[str, Any] = int(config['seed'] ) _lowerCamelCase : Optional[int] = int(config['batch_size'] ) _lowerCamelCase : Dict = args.model_name_or_path set_seed(lowercase__ ) _lowerCamelCase, _lowerCamelCase : Optional[int] = get_dataloaders(lowercase__ , lowercase__ , lowercase__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _lowerCamelCase : int = AutoModelForSequenceClassification.from_pretrained(lowercase__ , return_dict=lowercase__ ) # Instantiate optimizer _lowerCamelCase : Optional[int] = ( AdamW if accelerator.state.deepspeed_plugin is None or 'optimizer' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) _lowerCamelCase : Union[str, Any] = optimizer_cls(params=model.parameters() , lr=lowercase__ ) if accelerator.state.deepspeed_plugin is not None: _lowerCamelCase : str = accelerator.state.deepspeed_plugin.deepspeed_config[ 'gradient_accumulation_steps' ] else: _lowerCamelCase : Tuple = 1 _lowerCamelCase : List[Any] = (len(lowercase__ ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): _lowerCamelCase : Tuple = get_linear_schedule_with_warmup( optimizer=lowercase__ , num_warmup_steps=0 , num_training_steps=lowercase__ , ) else: _lowerCamelCase : Any = DummyScheduler(lowercase__ , total_num_steps=lowercase__ , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : Union[str, Any] = accelerator.prepare( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) # We need to keep track of how many total steps we have iterated over _lowerCamelCase : Union[str, Any] = 0 # We also need to keep track of the stating epoch so files are named properly _lowerCamelCase : Dict = 0 # Now we train the model _lowerCamelCase : Dict = evaluate.load('glue' , 'mrpc' ) _lowerCamelCase : Optional[int] = 0 _lowerCamelCase : str = {} for epoch in range(lowercase__ , lowercase__ ): model.train() for step, batch in enumerate(lowercase__ ): _lowerCamelCase : List[Any] = model(**lowercase__ ) _lowerCamelCase : int = outputs.loss _lowerCamelCase : Dict = loss / gradient_accumulation_steps accelerator.backward(lowercase__ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 model.eval() _lowerCamelCase : Union[str, Any] = 0 for step, batch in enumerate(lowercase__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _lowerCamelCase : Optional[int] = model(**lowercase__ ) _lowerCamelCase : Dict = outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times _lowerCamelCase, _lowerCamelCase : List[str] = accelerator.gather( (predictions, batch['labels']) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(lowercase__ ) - 1: _lowerCamelCase : Optional[Any] = predictions[: len(eval_dataloader.dataset ) - samples_seen] _lowerCamelCase : Dict = references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=lowercase__ , references=lowercase__ , ) _lowerCamelCase : List[Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f'''epoch {epoch}:''' , lowercase__ ) _lowerCamelCase : Tuple = eval_metric['accuracy'] if best_performance < eval_metric["accuracy"]: _lowerCamelCase : str = eval_metric['accuracy'] if args.performance_lower_bound is not None: assert ( args.performance_lower_bound <= best_performance ), f'''Best performance metric {best_performance} is lower than the lower bound {args.performance_lower_bound}''' accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , 'all_results.json' ) , 'w' ) as f: json.dump(lowercase__ , lowercase__ ) def _snake_case ( ): _lowerCamelCase : Any = argparse.ArgumentParser(description='Simple example of training script tracking peak GPU memory usage.' ) parser.add_argument( '--model_name_or_path' , type=lowercase__ , default='bert-base-cased' , help='Path to pretrained model or model identifier from huggingface.co/models.' , required=lowercase__ , ) parser.add_argument( '--output_dir' , type=lowercase__ , default='.' , help='Optional save directory where all checkpoint folders will be stored. Default is the current working directory.' , ) parser.add_argument( '--performance_lower_bound' , type=lowercase__ , default=lowercase__ , help='Optional lower bound for the performance metric. If set, the training will throw error when the performance metric drops below this value.' , ) parser.add_argument( '--num_epochs' , type=lowercase__ , default=3 , help='Number of train epochs.' , ) _lowerCamelCase : Optional[Any] = parser.parse_args() _lowerCamelCase : str = {'lr': 2E-5, 'num_epochs': args.num_epochs, 'seed': 42, 'batch_size': 16} training_function(lowercase__ , lowercase__ ) if __name__ == "__main__": main()
96
1
"""simple docstring""" def _snake_case ( ): return [ a * b * (1000 - a - b) for a in range(1 , 999 ) for b in range(lowercase__ , 999 ) if (a * a + b * b == (1000 - a - b) ** 2) ][0] if __name__ == "__main__": print(F"{solution() = }")
96
"""simple docstring""" from __future__ import annotations import copy import tempfile import unittest from transformers import CONFIG_MAPPING, AutoConfig, BertConfig, GPTaConfig, TaConfig, TapasConfig, is_tf_available from transformers.testing_utils import ( DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tensorflow_probability, require_tf, slow, ) from ..bert.test_modeling_bert import BertModelTester if is_tf_available(): from transformers import ( TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelForTableQuestionAnswering, TFAutoModelForTokenClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFFunnelBaseModel, TFFunnelModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, TFTapasForQuestionAnswering, ) from transformers.models.auto.modeling_tf_auto import ( TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_MAPPING, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.tapas.modeling_tf_tapas import TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """new-model""" if is_tf_available(): class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = NewModelConfig @require_tf class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' @slow def A_ ( self ): _lowerCamelCase : List[str] = 'bert-base-cased' _lowerCamelCase : Union[str, Any] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Union[str, Any] = TFAutoModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): _lowerCamelCase : List[str] = 'bert-base-cased' _lowerCamelCase : int = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : int = TFAutoModelForPreTraining.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Dict = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : int = TFAutoModelForCausalLM.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : str = TFAutoModelForCausalLM.from_pretrained(lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : List[Any] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : str = TFAutoModelWithLMHead.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Tuple = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Optional[int] = TFAutoModelForMaskedLM.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : Tuple = TFAutoModelForMaskedLM.from_pretrained(lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Optional[int] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Tuple = TFAutoModelForSeqaSeqLM.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : Tuple = TFAutoModelForSeqaSeqLM.from_pretrained(lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: _lowerCamelCase : str = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Union[str, Any] = TFAutoModelForSequenceClassification.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: _lowerCamelCase : Optional[Any] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : List[str] = TFAutoModelForQuestionAnswering.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow @require_tensorflow_probability def A_ ( self ): for model_name in TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST[5:6]: _lowerCamelCase : Dict = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Dict = TFAutoModelForTableQuestionAnswering.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : List[Any] = TFAutoModelForTableQuestionAnswering.from_pretrained( lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) def A_ ( self ): _lowerCamelCase : int = TFAutoModelWithLMHead.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowercase ) , 14410 ) def A_ ( self ): _lowerCamelCase : Any = TFAutoModelWithLMHead.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowercase ) , 14410 ) def A_ ( self ): # For the auto model mapping, FunnelConfig has two models: FunnelModel and FunnelBaseModel _lowerCamelCase : List[str] = TFAutoModel.from_pretrained('sgugger/funnel-random-tiny' ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Optional[int] = copy.deepcopy(model.config ) _lowerCamelCase : Dict = ['FunnelBaseModel'] _lowerCamelCase : List[Any] = TFAutoModel.from_config(lowercase ) self.assertIsInstance(lowercase , lowercase ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowercase ) _lowerCamelCase : Tuple = TFAutoModel.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) def A_ ( self ): try: AutoConfig.register('new-model' , lowercase ) _lowerCamelCase : Tuple = [ TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSequenceClassification, TFAutoModelForTokenClassification, ] for auto_class in auto_classes: with self.subTest(auto_class.__name__ ): # Wrong config class will raise an error with self.assertRaises(lowercase ): auto_class.register(lowercase , lowercase ) auto_class.register(lowercase , lowercase ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(lowercase ): auto_class.register(lowercase , lowercase ) # Now that the config is registered, it can be used as any other config with the auto-API _lowerCamelCase : Optional[Any] = BertModelTester(self ).get_config() _lowerCamelCase : Dict = NewModelConfig(**tiny_config.to_dict() ) _lowerCamelCase : int = auto_class.from_config(lowercase ) self.assertIsInstance(lowercase , lowercase ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowercase ) _lowerCamelCase : List[Any] = auto_class.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"] for mapping in ( TF_MODEL_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, ): if NewModelConfig in mapping._extra_content: del mapping._extra_content[NewModelConfig] def A_ ( self ): with self.assertRaisesRegex( lowercase , 'bert-base is not a local folder and is not a valid model identifier' ): _lowerCamelCase : Union[str, Any] = TFAutoModel.from_pretrained('bert-base' ) def A_ ( self ): with self.assertRaisesRegex( lowercase , r'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): _lowerCamelCase : str = TFAutoModel.from_pretrained(lowercase , revision='aaaaaa' ) def A_ ( self ): with self.assertRaisesRegex( lowercase , 'hf-internal-testing/config-no-model does not appear to have a file named pytorch_model.bin' , ): _lowerCamelCase : Tuple = TFAutoModel.from_pretrained('hf-internal-testing/config-no-model' ) def A_ ( self ): with self.assertRaisesRegex(lowercase , 'Use `from_pt=True` to load this model' ): _lowerCamelCase : Tuple = TFAutoModel.from_pretrained('hf-internal-testing/tiny-bert-pt-only' ) def A_ ( self ): # Make sure we have cached the model. _lowerCamelCase : Optional[int] = TFAutoModel.from_pretrained('hf-internal-testing/tiny-random-bert' ) with RequestCounter() as counter: _lowerCamelCase : Optional[int] = TFAutoModel.from_pretrained('hf-internal-testing/tiny-random-bert' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 ) # With a sharded checkpoint _lowerCamelCase : int = TFAutoModel.from_pretrained('ArthurZ/tiny-random-bert-sharded' ) with RequestCounter() as counter: _lowerCamelCase : List[Any] = TFAutoModel.from_pretrained('ArthurZ/tiny-random-bert-sharded' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 )
96
1
"""simple docstring""" import unittest import torch from diffusers import VQModel from diffusers.utils import floats_tensor, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class lowerCAmelCase__ ( lowercase, lowercase, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = VQModel lowerCamelCase__ = """sample""" @property def A_ ( self , lowercase=(32, 32) ): _lowerCamelCase : str = 4 _lowerCamelCase : Tuple = 3 _lowerCamelCase : Any = floats_tensor((batch_size, num_channels) + sizes ).to(lowercase ) return {"sample": image} @property def A_ ( self ): return (3, 32, 32) @property def A_ ( self ): return (3, 32, 32) def A_ ( self ): _lowerCamelCase : List[str] = { 'block_out_channels': [32, 64], 'in_channels': 3, 'out_channels': 3, 'down_block_types': ['DownEncoderBlock2D', 'DownEncoderBlock2D'], 'up_block_types': ['UpDecoderBlock2D', 'UpDecoderBlock2D'], 'latent_channels': 3, } _lowerCamelCase : List[str] = self.dummy_input return init_dict, inputs_dict def A_ ( self ): pass def A_ ( self ): pass def A_ ( self ): _lowerCamelCase, _lowerCamelCase : List[str] = VQModel.from_pretrained('fusing/vqgan-dummy' , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertEqual(len(loading_info['missing_keys'] ) , 0 ) model.to(lowercase ) _lowerCamelCase : Dict = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def A_ ( self ): _lowerCamelCase : Union[str, Any] = VQModel.from_pretrained('fusing/vqgan-dummy' ) model.to(lowercase ).eval() torch.manual_seed(0 ) if torch.cuda.is_available(): torch.cuda.manual_seed_all(0 ) _lowerCamelCase : Any = torch.randn(1 , model.config.in_channels , model.config.sample_size , model.config.sample_size ) _lowerCamelCase : List[str] = image.to(lowercase ) with torch.no_grad(): _lowerCamelCase : int = model(lowercase ).sample _lowerCamelCase : List[str] = output[0, -1, -3:, -3:].flatten().cpu() # fmt: off _lowerCamelCase : Tuple = torch.tensor([-0.01_53, -0.40_44, -0.18_80, -0.51_61, -0.24_18, -0.40_72, -0.16_12, -0.06_33, -0.01_43] ) # fmt: on self.assertTrue(torch.allclose(lowercase , lowercase , atol=1E-3 ) )
96
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ = {"""configuration_ibert""": ["""IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """IBertConfig""", """IBertOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ """IBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """IBertForMaskedLM""", """IBertForMultipleChoice""", """IBertForQuestionAnswering""", """IBertForSequenceClassification""", """IBertForTokenClassification""", """IBertModel""", """IBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
96
1
"""simple docstring""" from __future__ import annotations from typing import Any class lowerCAmelCase__ : '''simple docstring''' def __init__( self , lowercase , lowercase , lowercase = 0 ): _lowerCamelCase, _lowerCamelCase : Tuple = row, column _lowerCamelCase : int = [[default_value for c in range(lowercase )] for r in range(lowercase )] def __str__( self ): _lowerCamelCase : Tuple = F'''Matrix consist of {self.row} rows and {self.column} columns\n''' # Make string identifier _lowerCamelCase : Optional[int] = 0 for row_vector in self.array: for obj in row_vector: _lowerCamelCase : int = max(lowercase , len(str(lowercase ) ) ) _lowerCamelCase : Optional[int] = F'''%{max_element_length}s''' # Make string and return def single_line(lowercase ) -> str: nonlocal string_format_identifier _lowerCamelCase : int = '[' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(lowercase ) for row_vector in self.array ) return s def __repr__( self ): return str(self ) def A_ ( self , lowercase ): if not (isinstance(lowercase , (list, tuple) ) and len(lowercase ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self , lowercase ): assert self.validate_indicies(lowercase ) return self.array[loc[0]][loc[1]] def __setitem__( self , lowercase , lowercase ): assert self.validate_indicies(lowercase ) _lowerCamelCase : Dict = value def __add__( self , lowercase ): assert isinstance(lowercase , lowercase ) assert self.row == another.row and self.column == another.column # Add _lowerCamelCase : Union[str, Any] = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): _lowerCamelCase : List[str] = self[r, c] + another[r, c] return result def __neg__( self ): _lowerCamelCase : int = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): _lowerCamelCase : List[str] = -self[r, c] return result def __sub__( self , lowercase ): return self + (-another) def __mul__( self , lowercase ): if isinstance(lowercase , (int, float) ): # Scalar multiplication _lowerCamelCase : int = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): _lowerCamelCase : Optional[Any] = self[r, c] * another return result elif isinstance(lowercase , lowercase ): # Matrix multiplication assert self.column == another.row _lowerCamelCase : Dict = Matrix(self.row , another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: _lowerCamelCase : Optional[int] = F'''Unsupported type given for another ({type(lowercase )})''' raise TypeError(lowercase ) def A_ ( self ): _lowerCamelCase : Any = Matrix(self.column , self.row ) for r in range(self.row ): for c in range(self.column ): _lowerCamelCase : Optional[Any] = self[r, c] return result def A_ ( self , lowercase , lowercase ): assert isinstance(lowercase , lowercase ) and isinstance(lowercase , lowercase ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate _lowerCamelCase : Optional[int] = v.transpose() _lowerCamelCase : Tuple = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def _snake_case ( ): # a^(-1) _lowerCamelCase : Optional[Any] = Matrix(3 , 3 , 0 ) for i in range(3 ): _lowerCamelCase : Optional[Any] = 1 print(f'''a^(-1) is {ainv}''' ) # u, v _lowerCamelCase : Tuple = Matrix(3 , 1 , 0 ) _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : Optional[int] = 1, 2, -3 _lowerCamelCase : Union[str, Any] = Matrix(3 , 1 , 0 ) _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : Any = 4, -2, 5 print(f'''u is {u}''' ) print(f'''v is {v}''' ) print(f'''uv^T is {u * v.transpose()}''' ) # Sherman Morrison print(f'''(a + uv^T)^(-1) is {ainv.sherman_morrison(lowercase__ , lowercase__ )}''' ) def _snake_case ( ): import doctest doctest.testmod() testa()
96
"""simple docstring""" import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : Tuple = f'''{sampling_rate}''' _lowerCamelCase : str = '1' _lowerCamelCase : str = 'f32le' _lowerCamelCase : Union[str, Any] = [ 'ffmpeg', '-i', 'pipe:0', '-ac', ac, '-ar', ar, '-f', format_for_conversion, '-hide_banner', '-loglevel', 'quiet', 'pipe:1', ] try: with subprocess.Popen(lowercase__ , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: _lowerCamelCase : str = ffmpeg_process.communicate(lowercase__ ) except FileNotFoundError as error: raise ValueError('ffmpeg was not found but is required to load audio files from filename' ) from error _lowerCamelCase : List[Any] = output_stream[0] _lowerCamelCase : Tuple = np.frombuffer(lowercase__ , np.floataa ) if audio.shape[0] == 0: raise ValueError('Malformed soundfile' ) return audio def _snake_case ( lowercase__ , lowercase__ , lowercase__ = "f32le" , ): _lowerCamelCase : Optional[Any] = f'''{sampling_rate}''' _lowerCamelCase : List[str] = '1' if format_for_conversion == "s16le": _lowerCamelCase : List[str] = 2 elif format_for_conversion == "f32le": _lowerCamelCase : List[Any] = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) _lowerCamelCase : Dict = platform.system() if system == "Linux": _lowerCamelCase : Optional[int] = 'alsa' _lowerCamelCase : Optional[Any] = 'default' elif system == "Darwin": _lowerCamelCase : Optional[int] = 'avfoundation' _lowerCamelCase : Any = ':0' elif system == "Windows": _lowerCamelCase : Tuple = 'dshow' _lowerCamelCase : Tuple = 'default' _lowerCamelCase : Optional[int] = [ 'ffmpeg', '-f', format_, '-i', input_, '-ac', ac, '-ar', ar, '-f', format_for_conversion, '-fflags', 'nobuffer', '-hide_banner', '-loglevel', 'quiet', 'pipe:1', ] _lowerCamelCase : Tuple = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample _lowerCamelCase : List[Any] = _ffmpeg_stream(lowercase__ , lowercase__ ) for item in iterator: yield item def _snake_case ( lowercase__ , lowercase__ , lowercase__ = None , lowercase__ = None , lowercase__ = "f32le" , ): if stream_chunk_s is not None: _lowerCamelCase : int = stream_chunk_s else: _lowerCamelCase : Optional[Any] = chunk_length_s _lowerCamelCase : Optional[Any] = ffmpeg_microphone(lowercase__ , lowercase__ , format_for_conversion=lowercase__ ) if format_for_conversion == "s16le": _lowerCamelCase : List[str] = np.intaa _lowerCamelCase : str = 2 elif format_for_conversion == "f32le": _lowerCamelCase : Any = np.floataa _lowerCamelCase : List[Any] = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) if stride_length_s is None: _lowerCamelCase : Union[str, Any] = chunk_length_s / 6 _lowerCamelCase : Optional[int] = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(lowercase__ , (int, float) ): _lowerCamelCase : Any = [stride_length_s, stride_length_s] _lowerCamelCase : Tuple = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample _lowerCamelCase : Optional[Any] = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample _lowerCamelCase : List[Any] = datetime.datetime.now() _lowerCamelCase : Optional[int] = datetime.timedelta(seconds=lowercase__ ) for item in chunk_bytes_iter(lowercase__ , lowercase__ , stride=(stride_left, stride_right) , stream=lowercase__ ): # Put everything back in numpy scale _lowerCamelCase : List[Any] = np.frombuffer(item['raw'] , dtype=lowercase__ ) _lowerCamelCase : int = ( item['stride'][0] // size_of_sample, item['stride'][1] // size_of_sample, ) _lowerCamelCase : Optional[int] = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ = False ): _lowerCamelCase : int = B'' _lowerCamelCase, _lowerCamelCase : Dict = stride if stride_left + stride_right >= chunk_len: raise ValueError( f'''Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}''' ) _lowerCamelCase : str = 0 for raw in iterator: acc += raw if stream and len(lowercase__ ) < chunk_len: _lowerCamelCase : Optional[int] = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(lowercase__ ) >= chunk_len: # We are flushing the accumulator _lowerCamelCase : str = (_stride_left, stride_right) _lowerCamelCase : str = {'raw': acc[:chunk_len], 'stride': stride} if stream: _lowerCamelCase : List[Any] = False yield item _lowerCamelCase : Optional[Any] = stride_left _lowerCamelCase : str = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(lowercase__ ) > stride_left: _lowerCamelCase : Optional[Any] = {'raw': acc, 'stride': (_stride_left, 0)} if stream: _lowerCamelCase : Tuple = False yield item def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : int = 2**24 # 16Mo try: with subprocess.Popen(lowercase__ , stdout=subprocess.PIPE , bufsize=lowercase__ ) as ffmpeg_process: while True: _lowerCamelCase : Optional[Any] = ffmpeg_process.stdout.read(lowercase__ ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError('ffmpeg was not found but is required to stream audio files from filename' ) from error
96
1
"""simple docstring""" import functools from typing import Any def _snake_case ( lowercase__ , lowercase__ ): # Validation if not isinstance(lowercase__ , lowercase__ ) or len(lowercase__ ) == 0: raise ValueError('the string should be not empty string' ) if not isinstance(lowercase__ , lowercase__ ) or not all( isinstance(lowercase__ , lowercase__ ) and len(lowercase__ ) > 0 for item in words ): raise ValueError('the words should be a list of non-empty strings' ) # Build trie _lowerCamelCase : dict[str, Any] = {} _lowerCamelCase : List[Any] = 'WORD_KEEPER' for word in words: _lowerCamelCase : Dict = trie for c in word: if c not in trie_node: _lowerCamelCase : Any = {} _lowerCamelCase : str = trie_node[c] _lowerCamelCase : Optional[Any] = True _lowerCamelCase : Dict = len(lowercase__ ) # Dynamic programming method @functools.cache def is_breakable(lowercase__ ) -> bool: if index == len_string: return True _lowerCamelCase : List[Any] = trie for i in range(lowercase__ , lowercase__ ): _lowerCamelCase : Any = trie_node.get(string[i] , lowercase__ ) if trie_node is None: return False if trie_node.get(lowercase__ , lowercase__ ) and is_breakable(i + 1 ): return True return False return is_breakable(0 ) if __name__ == "__main__": import doctest doctest.testmod()
96
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = {"""ctrl""": """https://huggingface.co/ctrl/resolve/main/config.json"""} class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """ctrl""" lowerCamelCase__ = ["""past_key_values"""] lowerCamelCase__ = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowercase=246534 , lowercase=256 , lowercase=1280 , lowercase=8192 , lowercase=48 , lowercase=16 , lowercase=0.1 , lowercase=0.1 , lowercase=1E-6 , lowercase=0.02 , lowercase=True , **lowercase , ): _lowerCamelCase : Any = vocab_size _lowerCamelCase : Dict = n_positions _lowerCamelCase : Optional[int] = n_embd _lowerCamelCase : str = n_layer _lowerCamelCase : Union[str, Any] = n_head _lowerCamelCase : Any = dff _lowerCamelCase : int = resid_pdrop _lowerCamelCase : Dict = embd_pdrop _lowerCamelCase : Union[str, Any] = layer_norm_epsilon _lowerCamelCase : Tuple = initializer_range _lowerCamelCase : str = use_cache super().__init__(**lowercase )
96
1
"""simple docstring""" from PIL import Image def _snake_case ( lowercase__ ): _lowerCamelCase, _lowerCamelCase : Dict = image.size _lowerCamelCase : Optional[int] = 0 _lowerCamelCase : List[str] = image.load() for i in range(lowercase__ ): for j in range(lowercase__ ): _lowerCamelCase : List[Any] = pixels[j, i] mean += pixel mean //= width * height for j in range(lowercase__ ): for i in range(lowercase__ ): _lowerCamelCase : Optional[Any] = 255 if pixels[i, j] > mean else 0 return image if __name__ == "__main__": lowercase__ = mean_threshold(Image.open("""path_to_image""").convert("""L""")) image.save("""output_image_path""")
96
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator from typing import Any class lowerCAmelCase__ : '''simple docstring''' def __init__( self , lowercase ): _lowerCamelCase : Any = data _lowerCamelCase : Node | None = None class lowerCAmelCase__ : '''simple docstring''' def __init__( self ): _lowerCamelCase : str = None _lowerCamelCase : str = None def __iter__( self ): _lowerCamelCase : List[str] = self.head while self.head: yield node.data _lowerCamelCase : Optional[int] = node.next if node == self.head: break def __len__( self ): return sum(1 for _ in self ) def __repr__( self ): return "->".join(str(lowercase ) for item in iter(self ) ) def A_ ( self , lowercase ): self.insert_nth(len(self ) , lowercase ) def A_ ( self , lowercase ): self.insert_nth(0 , lowercase ) def A_ ( self , lowercase , lowercase ): if index < 0 or index > len(self ): raise IndexError('list index out of range.' ) _lowerCamelCase : List[Any] = Node(lowercase ) if self.head is None: _lowerCamelCase : str = new_node # first node points itself _lowerCamelCase : Union[str, Any] = new_node elif index == 0: # insert at head _lowerCamelCase : List[str] = self.head _lowerCamelCase : str = new_node else: _lowerCamelCase : Union[str, Any] = self.head for _ in range(index - 1 ): _lowerCamelCase : List[Any] = temp.next _lowerCamelCase : Union[str, Any] = temp.next _lowerCamelCase : List[str] = new_node if index == len(self ) - 1: # insert at tail _lowerCamelCase : Any = new_node def A_ ( self ): return self.delete_nth(0 ) def A_ ( self ): return self.delete_nth(len(self ) - 1 ) def A_ ( self , lowercase = 0 ): if not 0 <= index < len(self ): raise IndexError('list index out of range.' ) _lowerCamelCase : Any = self.head if self.head == self.tail: # just one node _lowerCamelCase : List[str] = None elif index == 0: # delete head node _lowerCamelCase : List[str] = self.tail.next.next _lowerCamelCase : Optional[int] = self.head.next else: _lowerCamelCase : Dict = self.head for _ in range(index - 1 ): _lowerCamelCase : List[Any] = temp.next _lowerCamelCase : int = temp.next _lowerCamelCase : Optional[int] = temp.next.next if index == len(self ) - 1: # delete at tail _lowerCamelCase : List[Any] = temp return delete_node.data def A_ ( self ): return len(self ) == 0 def _snake_case ( ): _lowerCamelCase : Union[str, Any] = CircularLinkedList() assert len(lowercase__ ) == 0 assert circular_linked_list.is_empty() is True assert str(lowercase__ ) == "" try: circular_linked_list.delete_front() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_tail() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_nth(-1 ) raise AssertionError except IndexError: assert True try: circular_linked_list.delete_nth(0 ) raise AssertionError except IndexError: assert True assert circular_linked_list.is_empty() is True for i in range(5 ): assert len(lowercase__ ) == i circular_linked_list.insert_nth(lowercase__ , i + 1 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 6 ) ) circular_linked_list.insert_tail(6 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 7 ) ) circular_linked_list.insert_head(0 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(0 , 7 ) ) assert circular_linked_list.delete_front() == 0 assert circular_linked_list.delete_tail() == 6 assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 6 ) ) assert circular_linked_list.delete_nth(2 ) == 3 circular_linked_list.insert_nth(2 , 3 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 6 ) ) assert circular_linked_list.is_empty() is False if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" import argparse from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection from diffusers import UnCLIPImageVariationPipeline, UnCLIPPipeline if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the output model.""") parser.add_argument( """--txt2img_unclip""", default="""kakaobrain/karlo-v1-alpha""", type=str, required=False, help="""The pretrained txt2img unclip.""", ) lowercase__ = parser.parse_args() lowercase__ = UnCLIPPipeline.from_pretrained(args.txtaimg_unclip) lowercase__ = CLIPImageProcessor() lowercase__ = CLIPVisionModelWithProjection.from_pretrained("""openai/clip-vit-large-patch14""") lowercase__ = UnCLIPImageVariationPipeline( decoder=txtaimg.decoder, text_encoder=txtaimg.text_encoder, tokenizer=txtaimg.tokenizer, text_proj=txtaimg.text_proj, feature_extractor=feature_extractor, image_encoder=image_encoder, super_res_first=txtaimg.super_res_first, super_res_last=txtaimg.super_res_last, decoder_scheduler=txtaimg.decoder_scheduler, super_res_scheduler=txtaimg.super_res_scheduler, ) imgaimg.save_pretrained(args.dump_path)
96
"""simple docstring""" # Lint as: python3 import os import re import urllib.parse from pathlib import Path from typing import Callable, List, Optional, Union from zipfile import ZipFile from ..utils.file_utils import cached_path, hf_github_url from ..utils.logging import get_logger from ..utils.version import Version lowercase__ = get_logger(__name__) class lowerCAmelCase__ : '''simple docstring''' lowerCamelCase__ = """dummy_data""" lowerCamelCase__ = """datasets""" lowerCamelCase__ = False def __init__( self , lowercase , lowercase , lowercase , lowercase = None , lowercase = False , lowercase = True , lowercase = None , ): _lowerCamelCase : Optional[Any] = 0 _lowerCamelCase : Dict = dataset_name _lowerCamelCase : Union[str, Any] = cache_dir _lowerCamelCase : Dict = use_local_dummy_data _lowerCamelCase : Tuple = config # download_callbacks take a single url as input _lowerCamelCase : List[Callable] = download_callbacks or [] # if False, it doesn't load existing files and it returns the paths of the dummy files relative # to the dummy_data zip file root _lowerCamelCase : Any = load_existing_dummy_data # TODO(PVP, QL) might need to make this more general _lowerCamelCase : str = str(lowercase ) # to be downloaded _lowerCamelCase : Union[str, Any] = None _lowerCamelCase : int = None @property def A_ ( self ): if self._dummy_file is None: _lowerCamelCase : Tuple = self.download_dummy_data() return self._dummy_file @property def A_ ( self ): if self.config is not None: # structure is dummy / config_name / version_name return os.path.join('dummy' , self.config.name , self.version_name ) # structure is dummy / version_name return os.path.join('dummy' , self.version_name ) @property def A_ ( self ): return os.path.join(self.dummy_data_folder , 'dummy_data.zip' ) def A_ ( self ): _lowerCamelCase : List[str] = ( self.local_path_to_dummy_data if self.use_local_dummy_data is True else self.github_path_to_dummy_data ) _lowerCamelCase : int = cached_path( lowercase , cache_dir=self.cache_dir , extract_compressed_file=lowercase , force_extract=lowercase ) return os.path.join(lowercase , self.dummy_file_name ) @property def A_ ( self ): return os.path.join(self.datasets_scripts_dir , self.dataset_name , self.dummy_zip_file ) @property def A_ ( self ): if self._bucket_url is None: _lowerCamelCase : List[Any] = hf_github_url(self.dataset_name , self.dummy_zip_file.replace(os.sep , '/' ) ) return self._bucket_url @property def A_ ( self ): # return full path if its a dir if os.path.isdir(self.dummy_file ): return self.dummy_file # else cut off path to file -> example `xsum`. return "/".join(self.dummy_file.replace(os.sep , '/' ).split('/' )[:-1] ) def A_ ( self , lowercase , *lowercase ): if self.load_existing_dummy_data: # dummy data is downloaded and tested _lowerCamelCase : Union[str, Any] = self.dummy_file else: # dummy data cannot be downloaded and only the path to dummy file is returned _lowerCamelCase : Union[str, Any] = self.dummy_file_name # special case when data_url is a dict if isinstance(lowercase , lowercase ): return self.create_dummy_data_dict(lowercase , lowercase ) elif isinstance(lowercase , (list, tuple) ): return self.create_dummy_data_list(lowercase , lowercase ) else: return self.create_dummy_data_single(lowercase , lowercase ) def A_ ( self , lowercase , *lowercase ): return self.download_and_extract(lowercase ) def A_ ( self , lowercase , lowercase ): return self.download_and_extract(lowercase ) def A_ ( self , lowercase , *lowercase , **lowercase ): return path def A_ ( self ): return {} def A_ ( self , lowercase , lowercase ): _lowerCamelCase : Optional[int] = {} for key, single_urls in data_url.items(): for download_callback in self.download_callbacks: if isinstance(lowercase , lowercase ): for single_url in single_urls: download_callback(lowercase ) else: _lowerCamelCase : List[Any] = single_urls download_callback(lowercase ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus if isinstance(lowercase , lowercase ): _lowerCamelCase : List[Any] = [os.path.join(lowercase , urllib.parse.quote_plus(Path(lowercase ).name ) ) for x in single_urls] else: _lowerCamelCase : Optional[int] = single_urls _lowerCamelCase : List[Any] = os.path.join(lowercase , urllib.parse.quote_plus(Path(lowercase ).name ) ) _lowerCamelCase : int = value # make sure that values are unique if all(isinstance(lowercase , lowercase ) for i in dummy_data_dict.values() ) and len(set(dummy_data_dict.values() ) ) < len( dummy_data_dict.values() ): # append key to value to make its name unique _lowerCamelCase : List[Any] = {key: value + key for key, value in dummy_data_dict.items()} return dummy_data_dict def A_ ( self , lowercase , lowercase ): _lowerCamelCase : Optional[Any] = [] # trick: if there are many shards named like `data.txt-000001-of-00300`, only use the first one _lowerCamelCase : List[str] = all(bool(re.findall('[0-9]{3,}-of-[0-9]{3,}' , lowercase ) ) for url in data_url ) _lowerCamelCase : int = all( url.startswith('https://ftp.ncbi.nlm.nih.gov/pubmed/baseline/pubmed' ) for url in data_url ) if data_url and (is_tf_records or is_pubmed_records): _lowerCamelCase : List[str] = [data_url[0]] * len(lowercase ) for single_url in data_url: for download_callback in self.download_callbacks: download_callback(lowercase ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus _lowerCamelCase : str = os.path.join(lowercase , urllib.parse.quote_plus(single_url.split('/' )[-1] ) ) dummy_data_list.append(lowercase ) return dummy_data_list def A_ ( self , lowercase , lowercase ): for download_callback in self.download_callbacks: download_callback(lowercase ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus _lowerCamelCase : Tuple = os.path.join(lowercase , urllib.parse.quote_plus(data_url.split('/' )[-1] ) ) if os.path.exists(lowercase ) or not self.load_existing_dummy_data: return value else: # Backward compatibility, maybe deprecate at one point. # For many datasets with single url calls to dl_manager.download_and_extract, # the dummy_data.zip file is actually the zipped downloaded file # while now we expected the dummy_data.zip file to be a directory containing # the downloaded file. return path_to_dummy_data def A_ ( self ): pass def A_ ( self ): pass def A_ ( self , lowercase ): def _iter_archive_members(lowercase ): # this preserves the order of the members inside the ZIP archive _lowerCamelCase : str = Path(self.dummy_file ).parent _lowerCamelCase : Union[str, Any] = path.relative_to(lowercase ) with ZipFile(self.local_path_to_dummy_data ) as zip_file: _lowerCamelCase : List[str] = zip_file.namelist() for member in members: if member.startswith(relative_path.as_posix() ): yield dummy_parent_path.joinpath(lowercase ) _lowerCamelCase : Optional[int] = Path(lowercase ) _lowerCamelCase : Dict = _iter_archive_members(lowercase ) if self.use_local_dummy_data else path.rglob('*' ) for file_path in file_paths: if file_path.is_file() and not file_path.name.startswith(('.', '__') ): yield file_path.relative_to(lowercase ).as_posix(), file_path.open('rb' ) def A_ ( self , lowercase ): if not isinstance(lowercase , lowercase ): _lowerCamelCase : List[str] = [paths] for path in paths: if os.path.isfile(lowercase ): if os.path.basename(lowercase ).startswith(('.', '__') ): return yield path else: for dirpath, dirnames, filenames in os.walk(lowercase ): if os.path.basename(lowercase ).startswith(('.', '__') ): continue dirnames.sort() for filename in sorted(lowercase ): if filename.startswith(('.', '__') ): continue yield os.path.join(lowercase , lowercase )
96
1
"""simple docstring""" import copy from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """encoder-decoder""" lowerCamelCase__ = True def __init__( self , **lowercase ): super().__init__(**lowercase ) assert ( "encoder" in kwargs and "decoder" in kwargs ), "Config has to be initialized with encoder and decoder config" _lowerCamelCase : int = kwargs.pop('encoder' ) _lowerCamelCase : Tuple = encoder_config.pop('model_type' ) _lowerCamelCase : int = kwargs.pop('decoder' ) _lowerCamelCase : List[str] = decoder_config.pop('model_type' ) from ..auto.configuration_auto import AutoConfig _lowerCamelCase : Optional[int] = AutoConfig.for_model(lowercase , **lowercase ) _lowerCamelCase : Tuple = AutoConfig.for_model(lowercase , **lowercase ) _lowerCamelCase : Dict = True @classmethod def A_ ( cls , lowercase , lowercase , **lowercase ): logger.info('Set `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config' ) _lowerCamelCase : Optional[Any] = True _lowerCamelCase : Dict = True return cls(encoder=encoder_config.to_dict() , decoder=decoder_config.to_dict() , **lowercase ) def A_ ( self ): _lowerCamelCase : Optional[Any] = copy.deepcopy(self.__dict__ ) _lowerCamelCase : Dict = self.encoder.to_dict() _lowerCamelCase : Optional[Any] = self.decoder.to_dict() _lowerCamelCase : List[str] = self.__class__.model_type return output
96
"""simple docstring""" def _snake_case ( lowercase__ ): stooge(lowercase__ , 0 , len(lowercase__ ) - 1 ) return arr def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: _lowerCamelCase, _lowerCamelCase : Optional[Any] = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: _lowerCamelCase : Union[str, Any] = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(lowercase__ , lowercase__ , (h - t) ) # Recursively sort last 2/3 elements stooge(lowercase__ , i + t , (lowercase__) ) # Recursively sort first 2/3 elements stooge(lowercase__ , lowercase__ , (h - t) ) if __name__ == "__main__": lowercase__ = input("""Enter numbers separated by a comma:\n""").strip() lowercase__ = [int(item) for item in user_input.split(""",""")] print(stooge_sort(unsorted))
96
1
"""simple docstring""" # This code is adapted from OpenAI's release # https://github.com/openai/human-eval/blob/master/human_eval/execution.py import contextlib import faulthandler import io import multiprocessing import os import platform import signal import tempfile def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : Any = multiprocessing.Manager() _lowerCamelCase : str = manager.list() _lowerCamelCase : int = multiprocessing.Process(target=lowercase__ , args=(check_program, result, timeout) ) p.start() p.join(timeout=timeout + 1 ) if p.is_alive(): p.kill() if not result: result.append('timed out' ) return { "task_id": task_id, "passed": result[0] == "passed", "result": result[0], "completion_id": completion_id, } def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): with create_tempdir(): # These system calls are needed when cleaning up tempdir. import os import shutil _lowerCamelCase : Optional[Any] = shutil.rmtree _lowerCamelCase : Dict = os.rmdir _lowerCamelCase : int = os.chdir # Disable functionalities that can make destructive changes to the test. reliability_guard() # Run program. try: _lowerCamelCase : Dict = {} with swallow_io(): with time_limit(lowercase__ ): exec(lowercase__ , lowercase__ ) result.append('passed' ) except TimeoutException: result.append('timed out' ) except BaseException as e: result.append(f'''failed: {e}''' ) # Needed for cleaning up. _lowerCamelCase : int = rmtree _lowerCamelCase : Tuple = rmdir _lowerCamelCase : List[Any] = chdir @contextlib.contextmanager def _snake_case ( lowercase__ ): def signal_handler(lowercase__ , lowercase__ ): raise TimeoutException('Timed out!' ) signal.setitimer(signal.ITIMER_REAL , lowercase__ ) signal.signal(signal.SIGALRM , lowercase__ ) try: yield finally: signal.setitimer(signal.ITIMER_REAL , 0 ) @contextlib.contextmanager def _snake_case ( ): _lowerCamelCase : Tuple = WriteOnlyStringIO() with contextlib.redirect_stdout(lowercase__ ): with contextlib.redirect_stderr(lowercase__ ): with redirect_stdin(lowercase__ ): yield @contextlib.contextmanager def _snake_case ( ): with tempfile.TemporaryDirectory() as dirname: with chdir(lowercase__ ): yield dirname class lowerCAmelCase__ ( lowercase ): '''simple docstring''' pass class lowerCAmelCase__ ( io.StringIO ): '''simple docstring''' def A_ ( self , *lowercase , **lowercase ): raise OSError def A_ ( self , *lowercase , **lowercase ): raise OSError def A_ ( self , *lowercase , **lowercase ): raise OSError def A_ ( self , *lowercase , **lowercase ): return False class lowerCAmelCase__ ( contextlib._RedirectStream ): # type: ignore '''simple docstring''' lowerCamelCase__ = """stdin""" @contextlib.contextmanager def _snake_case ( lowercase__ ): if root == ".": yield return _lowerCamelCase : List[str] = os.getcwd() os.chdir(lowercase__ ) try: yield except BaseException as exc: raise exc finally: os.chdir(lowercase__ ) def _snake_case ( lowercase__=None ): if maximum_memory_bytes is not None: import resource resource.setrlimit(resource.RLIMIT_AS , (maximum_memory_bytes, maximum_memory_bytes) ) resource.setrlimit(resource.RLIMIT_DATA , (maximum_memory_bytes, maximum_memory_bytes) ) if not platform.uname().system == "Darwin": resource.setrlimit(resource.RLIMIT_STACK , (maximum_memory_bytes, maximum_memory_bytes) ) faulthandler.disable() import builtins _lowerCamelCase : Dict = None _lowerCamelCase : Any = None import os _lowerCamelCase : Optional[Any] = '1' _lowerCamelCase : Union[str, Any] = None _lowerCamelCase : Dict = None _lowerCamelCase : str = None _lowerCamelCase : Optional[Any] = None _lowerCamelCase : Optional[Any] = None _lowerCamelCase : Union[str, Any] = None _lowerCamelCase : List[str] = None _lowerCamelCase : Any = None _lowerCamelCase : str = None _lowerCamelCase : int = None _lowerCamelCase : int = None _lowerCamelCase : Tuple = None _lowerCamelCase : Any = None _lowerCamelCase : List[Any] = None _lowerCamelCase : Any = None _lowerCamelCase : Any = None _lowerCamelCase : List[str] = None _lowerCamelCase : Optional[int] = None _lowerCamelCase : Any = None _lowerCamelCase : Any = None _lowerCamelCase : List[Any] = None _lowerCamelCase : List[str] = None _lowerCamelCase : Optional[Any] = None _lowerCamelCase : Union[str, Any] = None _lowerCamelCase : List[Any] = None _lowerCamelCase : Optional[int] = None _lowerCamelCase : str = None import shutil _lowerCamelCase : str = None _lowerCamelCase : int = None _lowerCamelCase : Any = None import subprocess _lowerCamelCase : List[Any] = None # type: ignore _lowerCamelCase : Union[str, Any] = None import sys _lowerCamelCase : Optional[int] = None _lowerCamelCase : Dict = None _lowerCamelCase : str = None _lowerCamelCase : List[Any] = None _lowerCamelCase : int = None
96
"""simple docstring""" import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = ["""image_processor""", """tokenizer"""] lowerCamelCase__ = """BlipImageProcessor""" lowerCamelCase__ = """AutoTokenizer""" def __init__( self , lowercase , lowercase , lowercase ): super().__init__(lowercase , lowercase ) # add QFormer tokenizer _lowerCamelCase : int = qformer_tokenizer def __call__( self , lowercase = None , lowercase = None , lowercase = True , lowercase = False , lowercase = None , lowercase = None , lowercase = 0 , lowercase = None , lowercase = None , lowercase = False , lowercase = False , lowercase = False , lowercase = False , lowercase = False , lowercase = True , lowercase = None , **lowercase , ): if images is None and text is None: raise ValueError('You have to specify at least images or text.' ) _lowerCamelCase : int = BatchFeature() if text is not None: _lowerCamelCase : List[str] = self.tokenizer( text=lowercase , add_special_tokens=lowercase , padding=lowercase , truncation=lowercase , max_length=lowercase , stride=lowercase , pad_to_multiple_of=lowercase , return_attention_mask=lowercase , return_overflowing_tokens=lowercase , return_special_tokens_mask=lowercase , return_offsets_mapping=lowercase , return_token_type_ids=lowercase , return_length=lowercase , verbose=lowercase , return_tensors=lowercase , **lowercase , ) encoding.update(lowercase ) _lowerCamelCase : List[str] = self.qformer_tokenizer( text=lowercase , add_special_tokens=lowercase , padding=lowercase , truncation=lowercase , max_length=lowercase , stride=lowercase , pad_to_multiple_of=lowercase , return_attention_mask=lowercase , return_overflowing_tokens=lowercase , return_special_tokens_mask=lowercase , return_offsets_mapping=lowercase , return_token_type_ids=lowercase , return_length=lowercase , verbose=lowercase , return_tensors=lowercase , **lowercase , ) _lowerCamelCase : List[Any] = qformer_text_encoding.pop('input_ids' ) _lowerCamelCase : Tuple = qformer_text_encoding.pop('attention_mask' ) if images is not None: _lowerCamelCase : int = self.image_processor(lowercase , return_tensors=lowercase ) encoding.update(lowercase ) return encoding def A_ ( self , *lowercase , **lowercase ): return self.tokenizer.batch_decode(*lowercase , **lowercase ) def A_ ( self , *lowercase , **lowercase ): return self.tokenizer.decode(*lowercase , **lowercase ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def A_ ( self ): _lowerCamelCase : Union[str, Any] = self.tokenizer.model_input_names _lowerCamelCase : Any = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def A_ ( self , lowercase , **lowercase ): if os.path.isfile(lowercase ): raise ValueError(F'''Provided path ({save_directory}) should be a directory, not a file''' ) os.makedirs(lowercase , exist_ok=lowercase ) _lowerCamelCase : Optional[Any] = os.path.join(lowercase , 'qformer_tokenizer' ) self.qformer_tokenizer.save_pretrained(lowercase ) return super().save_pretrained(lowercase , **lowercase ) @classmethod def A_ ( cls , lowercase , **lowercase ): _lowerCamelCase : Union[str, Any] = AutoTokenizer.from_pretrained(lowercase , subfolder='qformer_tokenizer' ) _lowerCamelCase : Dict = cls._get_arguments_from_pretrained(lowercase , **lowercase ) args.append(lowercase ) return cls(*lowercase )
96
1
"""simple docstring""" from __future__ import annotations def _snake_case ( lowercase__ ): _lowerCamelCase : str = str(lowercase__ ) return n == n[::-1] def _snake_case ( lowercase__ = 1000000 ): _lowerCamelCase : Optional[int] = 0 for i in range(1 , lowercase__ ): if is_palindrome(lowercase__ ) and is_palindrome(bin(lowercase__ ).split('b' )[1] ): total += i return total if __name__ == "__main__": print(solution(int(str(input().strip()))))
96
"""simple docstring""" import json import logging import os import sys from time import time from unittest.mock import patch from transformers.testing_utils import TestCasePlus, require_torch_tpu logging.basicConfig(level=logging.DEBUG) lowercase__ = logging.getLogger() def _snake_case ( lowercase__ ): _lowerCamelCase : List[Any] = {} _lowerCamelCase : List[Any] = os.path.join(lowercase__ , 'all_results.json' ) if os.path.exists(lowercase__ ): with open(lowercase__ , 'r' ) as f: _lowerCamelCase : List[Any] = json.load(lowercase__ ) else: raise ValueError(f'''can\'t find {path}''' ) return results lowercase__ = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) @require_torch_tpu class lowerCAmelCase__ ( lowercase ): '''simple docstring''' def A_ ( self ): import xla_spawn _lowerCamelCase : List[Any] = self.get_auto_remove_tmp_dir() _lowerCamelCase : List[Any] = F''' ./examples/pytorch/text-classification/run_glue.py --num_cores=8 ./examples/pytorch/text-classification/run_glue.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --overwrite_output_dir --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --do_train --do_eval --debug tpu_metrics_debug --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --max_steps=10 --warmup_steps=2 --seed=42 --max_seq_length=128 '''.split() with patch.object(lowercase , 'argv' , lowercase ): _lowerCamelCase : Dict = time() xla_spawn.main() _lowerCamelCase : Any = time() _lowerCamelCase : Optional[int] = get_results(lowercase ) self.assertGreaterEqual(result['eval_accuracy'] , 0.75 ) # Assert that the script takes less than 500 seconds to make sure it doesn't hang. self.assertLess(end - start , 500 ) def A_ ( self ): import xla_spawn _lowerCamelCase : Tuple = '\n ./tests/test_trainer_tpu.py\n --num_cores=8\n ./tests/test_trainer_tpu.py\n '.split() with patch.object(lowercase , 'argv' , lowercase ): xla_spawn.main()
96
1
"""simple docstring""" import json import logging import os import sys from time import time from unittest.mock import patch from transformers.testing_utils import TestCasePlus, require_torch_tpu logging.basicConfig(level=logging.DEBUG) lowercase__ = logging.getLogger() def _snake_case ( lowercase__ ): _lowerCamelCase : List[Any] = {} _lowerCamelCase : List[Any] = os.path.join(lowercase__ , 'all_results.json' ) if os.path.exists(lowercase__ ): with open(lowercase__ , 'r' ) as f: _lowerCamelCase : List[Any] = json.load(lowercase__ ) else: raise ValueError(f'''can\'t find {path}''' ) return results lowercase__ = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) @require_torch_tpu class lowerCAmelCase__ ( lowercase ): '''simple docstring''' def A_ ( self ): import xla_spawn _lowerCamelCase : List[Any] = self.get_auto_remove_tmp_dir() _lowerCamelCase : List[Any] = F''' ./examples/pytorch/text-classification/run_glue.py --num_cores=8 ./examples/pytorch/text-classification/run_glue.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --overwrite_output_dir --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --do_train --do_eval --debug tpu_metrics_debug --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --max_steps=10 --warmup_steps=2 --seed=42 --max_seq_length=128 '''.split() with patch.object(lowercase , 'argv' , lowercase ): _lowerCamelCase : Dict = time() xla_spawn.main() _lowerCamelCase : Any = time() _lowerCamelCase : Optional[int] = get_results(lowercase ) self.assertGreaterEqual(result['eval_accuracy'] , 0.75 ) # Assert that the script takes less than 500 seconds to make sure it doesn't hang. self.assertLess(end - start , 500 ) def A_ ( self ): import xla_spawn _lowerCamelCase : Tuple = '\n ./tests/test_trainer_tpu.py\n --num_cores=8\n ./tests/test_trainer_tpu.py\n '.split() with patch.object(lowercase , 'argv' , lowercase ): xla_spawn.main()
96
"""simple docstring""" from __future__ import annotations import math import numpy as np from numpy.linalg import norm def _snake_case ( lowercase__ , lowercase__ ): return math.sqrt(sum(pow(a - b , 2 ) for a, b in zip(lowercase__ , lowercase__ ) ) ) def _snake_case ( lowercase__ , lowercase__ ): if dataset.ndim != value_array.ndim: _lowerCamelCase : Tuple = ( 'Wrong input data\'s dimensions... ' f'''dataset : {dataset.ndim}, value_array : {value_array.ndim}''' ) raise ValueError(lowercase__ ) try: if dataset.shape[1] != value_array.shape[1]: _lowerCamelCase : Optional[int] = ( 'Wrong input data\'s shape... ' f'''dataset : {dataset.shape[1]}, value_array : {value_array.shape[1]}''' ) raise ValueError(lowercase__ ) except IndexError: if dataset.ndim != value_array.ndim: raise TypeError('Wrong shape' ) if dataset.dtype != value_array.dtype: _lowerCamelCase : int = ( 'Input data have different datatype... ' f'''dataset : {dataset.dtype}, value_array : {value_array.dtype}''' ) raise TypeError(lowercase__ ) _lowerCamelCase : Optional[int] = [] for value in value_array: _lowerCamelCase : Tuple = euclidean(lowercase__ , dataset[0] ) _lowerCamelCase : Union[str, Any] = dataset[0].tolist() for dataset_value in dataset[1:]: _lowerCamelCase : Optional[Any] = euclidean(lowercase__ , lowercase__ ) if dist > temp_dist: _lowerCamelCase : List[Any] = temp_dist _lowerCamelCase : List[str] = dataset_value.tolist() answer.append([vector, dist] ) return answer def _snake_case ( lowercase__ , lowercase__ ): return np.dot(lowercase__ , lowercase__ ) / (norm(lowercase__ ) * norm(lowercase__ )) if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" import unittest from transformers import DonutProcessor lowercase__ = """naver-clova-ix/donut-base""" class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def A_ ( self ): _lowerCamelCase : Union[str, Any] = DonutProcessor.from_pretrained(lowercase ) def A_ ( self ): _lowerCamelCase : str = { 'name': 'John Doe', 'age': '99', 'city': 'Atlanta', 'state': 'GA', 'zip': '30301', 'phone': '123-4567', 'nicknames': [{'nickname': 'Johnny'}, {'nickname': 'JD'}], } _lowerCamelCase : int = ( '<s_name>John Doe</s_name><s_age>99</s_age><s_city>Atlanta</s_city>' '<s_state>GA</s_state><s_zip>30301</s_zip><s_phone>123-4567</s_phone>' '<s_nicknames><s_nickname>Johnny</s_nickname>' '<sep/><s_nickname>JD</s_nickname></s_nicknames>' ) _lowerCamelCase : str = self.processor.tokenajson(lowercase ) self.assertDictEqual(lowercase , lowercase )
96
"""simple docstring""" import socket def _snake_case ( ): _lowerCamelCase : List[Any] = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) _lowerCamelCase : Union[str, Any] = socket.gethostname() _lowerCamelCase : List[Any] = 12312 sock.connect((host, port) ) sock.send(B'Hello server!' ) with open('Received_file' , 'wb' ) as out_file: print('File opened' ) print('Receiving data...' ) while True: _lowerCamelCase : int = sock.recv(1024 ) if not data: break out_file.write(lowercase__ ) print('Successfully received the file' ) sock.close() print('Connection closed' ) if __name__ == "__main__": main()
96
1
"""simple docstring""" import unittest import numpy as np from diffusers import LMSDiscreteScheduler, OnnxStableDiffusionInpaintPipeline from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class lowerCAmelCase__ ( lowercase, unittest.TestCase ): '''simple docstring''' pass @nightly @require_onnxruntime @require_torch_gpu class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' @property def A_ ( self ): return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def A_ ( self ): _lowerCamelCase : List[Any] = ort.SessionOptions() _lowerCamelCase : Any = False return options def A_ ( self ): _lowerCamelCase : Optional[int] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/in_paint/overture-creations-5sI6fQgYIuo.png' ) _lowerCamelCase : List[str] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/in_paint/overture-creations-5sI6fQgYIuo_mask.png' ) _lowerCamelCase : Optional[int] = OnnxStableDiffusionInpaintPipeline.from_pretrained( 'runwayml/stable-diffusion-inpainting' , revision='onnx' , safety_checker=lowercase , feature_extractor=lowercase , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : Dict = 'A red cat sitting on a park bench' _lowerCamelCase : str = np.random.RandomState(0 ) _lowerCamelCase : Optional[Any] = pipe( prompt=lowercase , image=lowercase , mask_image=lowercase , guidance_scale=7.5 , num_inference_steps=10 , generator=lowercase , output_type='np' , ) _lowerCamelCase : Union[str, Any] = output.images _lowerCamelCase : List[str] = images[0, 255:258, 255:258, -1] assert images.shape == (1, 512, 512, 3) _lowerCamelCase : Any = np.array([0.25_14, 0.30_07, 0.35_17, 0.17_90, 0.23_82, 0.31_67, 0.19_44, 0.22_73, 0.24_64] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def A_ ( self ): _lowerCamelCase : int = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/in_paint/overture-creations-5sI6fQgYIuo.png' ) _lowerCamelCase : Tuple = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/in_paint/overture-creations-5sI6fQgYIuo_mask.png' ) _lowerCamelCase : str = LMSDiscreteScheduler.from_pretrained( 'runwayml/stable-diffusion-inpainting' , subfolder='scheduler' , revision='onnx' ) _lowerCamelCase : Any = OnnxStableDiffusionInpaintPipeline.from_pretrained( 'runwayml/stable-diffusion-inpainting' , revision='onnx' , scheduler=lowercase , safety_checker=lowercase , feature_extractor=lowercase , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : Dict = 'A red cat sitting on a park bench' _lowerCamelCase : str = np.random.RandomState(0 ) _lowerCamelCase : str = pipe( prompt=lowercase , image=lowercase , mask_image=lowercase , guidance_scale=7.5 , num_inference_steps=20 , generator=lowercase , output_type='np' , ) _lowerCamelCase : Any = output.images _lowerCamelCase : Union[str, Any] = images[0, 255:258, 255:258, -1] assert images.shape == (1, 512, 512, 3) _lowerCamelCase : int = np.array([0.00_86, 0.00_77, 0.00_83, 0.00_93, 0.01_07, 0.01_39, 0.00_94, 0.00_97, 0.01_25] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3
96
"""simple docstring""" from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record lowercase__ = """\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } """ lowercase__ = """\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. """ lowercase__ = """ Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for 'record': list of question-answer dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'prediction_text': the predicted answer text - for 'multirc': list of question-answer dictionaries with the following keys: - 'idx': index of the question-answer pair as specified by the dataset - 'prediction': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for 'record': list of question-answers dictionaries with the following keys: - 'idx': index of the question as specified by the dataset - 'answers': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for 'record': - 'exact_match': Exact match between answer and gold answer - 'f1': F1 score - for 'multirc': - 'exact_match': Exact match between answer and gold answer - 'f1_m': Per-question macro-F1 score - 'f1_a': Average F1 score over all answers - for 'axb': 'matthews_correlation': Matthew Correlation - for 'cb': - 'accuracy': Accuracy - 'f1': F1 score - for all others: - 'accuracy': Accuracy Examples: >>> super_glue_metric = datasets.load_metric('super_glue', 'copa') # any of [\"copa\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"boolq\", \"axg\"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'cb') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'record') >>> predictions = [{'idx': {'passage': 0, 'query': 0}, 'prediction_text': 'answer'}] >>> references = [{'idx': {'passage': 0, 'query': 0}, 'answers': ['answer', 'another_answer']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'multirc') >>> predictions = [{'idx': {'answer': 0, 'paragraph': 0, 'question': 0}, 'prediction': 0}, {'idx': {'answer': 1, 'paragraph': 2, 'question': 3}, 'prediction': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 1.0, 'f1_m': 1.0, 'f1_a': 1.0} >>> super_glue_metric = datasets.load_metric('super_glue', 'axb') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def _snake_case ( lowercase__ , lowercase__ ): return float((preds == labels).mean() ) def _snake_case ( lowercase__ , lowercase__ , lowercase__="binary" ): _lowerCamelCase : str = simple_accuracy(lowercase__ , lowercase__ ) _lowerCamelCase : Any = float(fa_score(y_true=lowercase__ , y_pred=lowercase__ , average=lowercase__ ) ) return { "accuracy": acc, "f1": fa, } def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : Any = {} for id_pred, label in zip(lowercase__ , lowercase__ ): _lowerCamelCase : Tuple = f'''{id_pred['idx']['paragraph']}-{id_pred['idx']['question']}''' _lowerCamelCase : Union[str, Any] = id_pred['prediction'] if question_id in question_map: question_map[question_id].append((pred, label) ) else: _lowerCamelCase : Optional[Any] = [(pred, label)] _lowerCamelCase, _lowerCamelCase : Optional[int] = [], [] for question, preds_labels in question_map.items(): _lowerCamelCase, _lowerCamelCase : Tuple = zip(*lowercase__ ) _lowerCamelCase : List[str] = fa_score(y_true=lowercase__ , y_pred=lowercase__ , average='macro' ) fas.append(lowercase__ ) _lowerCamelCase : int = int(sum(pred == label for pred, label in preds_labels ) == len(lowercase__ ) ) ems.append(lowercase__ ) _lowerCamelCase : Optional[Any] = float(sum(lowercase__ ) / len(lowercase__ ) ) _lowerCamelCase : Optional[int] = sum(lowercase__ ) / len(lowercase__ ) _lowerCamelCase : List[Any] = float(fa_score(y_true=lowercase__ , y_pred=[id_pred['prediction'] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCAmelCase__ ( datasets.Metric ): '''simple docstring''' def A_ ( self ): if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( 'You should supply a configuration name selected in ' '["boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg",]' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format='numpy' if not self.config_name == 'record' and not self.config_name == 'multirc' else None , ) def A_ ( self ): if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value('int64' ), "query": datasets.Value('int64' ), }, "prediction_text": datasets.Value('string' ), }, "references": { "idx": { "passage": datasets.Value('int64' ), "query": datasets.Value('int64' ), }, "answers": datasets.Sequence(datasets.Value('string' ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value('int64' ), "paragraph": datasets.Value('int64' ), "question": datasets.Value('int64' ), }, "prediction": datasets.Value('int64' ), }, "references": datasets.Value('int64' ), } else: return { "predictions": datasets.Value('int64' ), "references": datasets.Value('int64' ), } def A_ ( self , lowercase , lowercase ): if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(lowercase , lowercase )} elif self.config_name == "cb": return acc_and_fa(lowercase , lowercase , fa_avg='macro' ) elif self.config_name == "record": _lowerCamelCase : List[str] = [ { 'qas': [ {'id': ref['idx']['query'], 'answers': [{'text': ans} for ans in ref['answers']]} for ref in references ] } ] _lowerCamelCase : Union[str, Any] = {pred['idx']['query']: pred['prediction_text'] for pred in predictions} return evaluate_record(lowercase , lowercase )[0] elif self.config_name == "multirc": return evaluate_multirc(lowercase , lowercase ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(lowercase , lowercase )} else: raise KeyError( 'You should supply a configuration name selected in ' '["boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg",]' )
96
1
"""simple docstring""" lowercase__ = tuple[float, float, float] lowercase__ = tuple[float, float, float] def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : Dict = end_pointa[0] - end_pointa[0] _lowerCamelCase : List[Any] = end_pointa[1] - end_pointa[1] _lowerCamelCase : List[str] = end_pointa[2] - end_pointa[2] return (x, y, z) def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : Dict = ab[1] * ac[2] - ab[2] * ac[1] # *i _lowerCamelCase : Dict = (ab[0] * ac[2] - ab[2] * ac[0]) * -1 # *j _lowerCamelCase : List[str] = ab[0] * ac[1] - ab[1] * ac[0] # *k return (x, y, z) def _snake_case ( lowercase__ , lowercase__ ): return tuple(round(lowercase__ , lowercase__ ) for x in vector ) == (0, 0, 0) def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ = 10 ): _lowerCamelCase : Optional[int] = create_vector(lowercase__ , lowercase__ ) _lowerCamelCase : Optional[Any] = create_vector(lowercase__ , lowercase__ ) return is_zero_vector(get_ad_vectors_cross(lowercase__ , lowercase__ ) , lowercase__ )
96
"""simple docstring""" import unittest import numpy as np import torch from diffusers import DDIMPipeline, DDIMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow, torch_device from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class lowerCAmelCase__ ( lowercase, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = DDIMPipeline lowerCamelCase__ = UNCONDITIONAL_IMAGE_GENERATION_PARAMS lowerCamelCase__ = PipelineTesterMixin.required_optional_params - { """num_images_per_prompt""", """latents""", """callback""", """callback_steps""", } lowerCamelCase__ = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS lowerCamelCase__ = False def A_ ( self ): torch.manual_seed(0 ) _lowerCamelCase : List[Any] = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('DownBlock2D', 'AttnDownBlock2D') , up_block_types=('AttnUpBlock2D', 'UpBlock2D') , ) _lowerCamelCase : List[str] = DDIMScheduler() _lowerCamelCase : Optional[int] = {'unet': unet, 'scheduler': scheduler} return components def A_ ( self , lowercase , lowercase=0 ): if str(lowercase ).startswith('mps' ): _lowerCamelCase : Dict = torch.manual_seed(lowercase ) else: _lowerCamelCase : List[str] = torch.Generator(device=lowercase ).manual_seed(lowercase ) _lowerCamelCase : Tuple = { 'batch_size': 1, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs def A_ ( self ): _lowerCamelCase : Any = 'cpu' _lowerCamelCase : Tuple = self.get_dummy_components() _lowerCamelCase : Optional[Any] = self.pipeline_class(**lowercase ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : str = self.get_dummy_inputs(lowercase ) _lowerCamelCase : int = pipe(**lowercase ).images _lowerCamelCase : Any = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 32, 32, 3) ) _lowerCamelCase : Tuple = np.array( [1.000E00, 5.717E-01, 4.717E-01, 1.000E00, 0.000E00, 1.000E00, 3.000E-04, 0.000E00, 9.000E-04] ) _lowerCamelCase : str = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowercase , 1E-3 ) def A_ ( self ): super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) def A_ ( self ): super().test_save_load_local(expected_max_difference=3E-3 ) def A_ ( self ): super().test_save_load_optional_components(expected_max_difference=3E-3 ) def A_ ( self ): super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def A_ ( self ): _lowerCamelCase : Optional[Any] = 'google/ddpm-cifar10-32' _lowerCamelCase : Optional[Any] = UNetaDModel.from_pretrained(lowercase ) _lowerCamelCase : Dict = DDIMScheduler() _lowerCamelCase : Dict = DDIMPipeline(unet=lowercase , scheduler=lowercase ) ddim.to(lowercase ) ddim.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : List[str] = torch.manual_seed(0 ) _lowerCamelCase : str = ddim(generator=lowercase , eta=0.0 , output_type='numpy' ).images _lowerCamelCase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _lowerCamelCase : List[Any] = np.array([0.17_23, 0.16_17, 0.16_00, 0.16_26, 0.14_97, 0.15_13, 0.15_05, 0.14_42, 0.14_53] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def A_ ( self ): _lowerCamelCase : Optional[int] = 'google/ddpm-ema-bedroom-256' _lowerCamelCase : str = UNetaDModel.from_pretrained(lowercase ) _lowerCamelCase : str = DDIMScheduler.from_pretrained(lowercase ) _lowerCamelCase : Optional[int] = DDIMPipeline(unet=lowercase , scheduler=lowercase ) ddpm.to(lowercase ) ddpm.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : Tuple = torch.manual_seed(0 ) _lowerCamelCase : int = ddpm(generator=lowercase , output_type='numpy' ).images _lowerCamelCase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _lowerCamelCase : str = np.array([0.00_60, 0.02_01, 0.03_44, 0.00_24, 0.00_18, 0.00_02, 0.00_22, 0.00_00, 0.00_69] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
96
1
"""simple docstring""" from __future__ import annotations import collections import tempfile import unittest import numpy as np from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import is_tf_available, is_vision_available from ...test_modeling_tf_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_tf_bert import TFBertModelTester from ..clip.test_modeling_tf_clip import TFCLIPVisionModelTester from ..deit.test_modeling_tf_deit import TFDeiTModelTester from ..roberta.test_modeling_tf_roberta import TFRobertaModelTester from ..vit.test_modeling_tf_vit import TFViTModelTester if is_tf_available(): from transformers import ( TFBertModel, TFCLIPVisionModel, TFDeiTModel, TFRobertaModel, TFVisionTextDualEncoderModel, TFViTModel, VisionTextDualEncoderConfig, ) if is_vision_available(): from PIL import Image from transformers import VisionTextDualEncoderProcessor def _snake_case ( lowercase__ ): if isinstance(lowercase__ , collections.abc.Iterable ): return x return (x, x) @require_tf class lowerCAmelCase__ : '''simple docstring''' def A_ ( self , lowercase , lowercase ): pass def A_ ( self ): pass def A_ ( self ): pass def A_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase=None , **lowercase ): _lowerCamelCase : Tuple = VisionTextDualEncoderConfig.from_vision_text_configs(lowercase , lowercase ) _lowerCamelCase : int = TFVisionTextDualEncoderModel(lowercase ) _lowerCamelCase : Optional[Any] = model(input_ids=lowercase , pixel_values=lowercase , attention_mask=lowercase ) self.assertEqual(output['text_embeds'].shape , (input_ids.shape[0], config.projection_dim) ) self.assertEqual(output['image_embeds'].shape , (pixel_values.shape[0], config.projection_dim) ) def A_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase=None , **lowercase ): _lowerCamelCase, _lowerCamelCase : Dict = self.get_vision_text_model(lowercase , lowercase ) _lowerCamelCase : Optional[Any] = TFVisionTextDualEncoderModel(vision_model=lowercase , text_model=lowercase ) _lowerCamelCase : List[str] = model(input_ids=lowercase , pixel_values=lowercase , attention_mask=lowercase ) self.assertEqual(output['text_embeds'].shape , (input_ids.shape[0], model.config.projection_dim) ) self.assertEqual(output['image_embeds'].shape , (pixel_values.shape[0], model.config.projection_dim) ) def A_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase=None , **lowercase ): _lowerCamelCase, _lowerCamelCase : List[Any] = self.get_vision_text_model(lowercase , lowercase ) _lowerCamelCase : Tuple = {'vision_model': vision_model, 'text_model': text_model} _lowerCamelCase : Dict = TFVisionTextDualEncoderModel.from_vision_text_pretrained(**lowercase ) _lowerCamelCase : Dict = model(input_ids=lowercase , pixel_values=lowercase , attention_mask=lowercase ) self.assertEqual(output['text_embeds'].shape , (input_ids.shape[0], model.config.projection_dim) ) self.assertEqual(output['image_embeds'].shape , (pixel_values.shape[0], model.config.projection_dim) ) def A_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase=None , **lowercase ): _lowerCamelCase, _lowerCamelCase : int = self.get_vision_text_model(lowercase , lowercase ) _lowerCamelCase : List[Any] = TFVisionTextDualEncoderModel(vision_model=lowercase , text_model=lowercase ) _lowerCamelCase : Union[str, Any] = model(input_ids=lowercase , pixel_values=lowercase , attention_mask=lowercase ) _lowerCamelCase : Union[str, Any] = output[0].numpy() with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowercase ) _lowerCamelCase : List[Any] = TFVisionTextDualEncoderModel.from_pretrained(lowercase ) _lowerCamelCase : Tuple = model(input_ids=lowercase , pixel_values=lowercase , attention_mask=lowercase ) _lowerCamelCase : List[str] = after_output[0].numpy() _lowerCamelCase : Optional[int] = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowercase , 1E-5 ) def A_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase=None , **lowercase ): _lowerCamelCase, _lowerCamelCase : List[Any] = self.get_vision_text_model(lowercase , lowercase ) _lowerCamelCase : List[str] = TFVisionTextDualEncoderModel(vision_model=lowercase , text_model=lowercase ) _lowerCamelCase : Tuple = model( input_ids=lowercase , pixel_values=lowercase , attention_mask=lowercase , output_attentions=lowercase ) _lowerCamelCase : List[Any] = output.vision_model_output.attentions self.assertEqual(len(lowercase ) , vision_config.num_hidden_layers ) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) _lowerCamelCase : Any = to_atuple(vision_model.config.image_size ) _lowerCamelCase : List[Any] = to_atuple(vision_model.config.patch_size ) _lowerCamelCase : Optional[int] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) _lowerCamelCase : Tuple = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len) ) _lowerCamelCase : Union[str, Any] = output.text_model_output.attentions self.assertEqual(len(lowercase ) , text_config.num_hidden_layers ) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def A_ ( self , lowercase , lowercase , lowercase ): _lowerCamelCase : Tuple = np.abs((a - b) ).max() self.assertLessEqual(lowercase , lowercase , F'''Difference between torch and flax is {diff} (>= {tol}).''' ) def A_ ( self ): _lowerCamelCase : Any = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_model(**lowercase ) def A_ ( self ): _lowerCamelCase : Optional[Any] = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**lowercase ) def A_ ( self ): _lowerCamelCase : Optional[int] = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**lowercase ) def A_ ( self ): _lowerCamelCase : List[Any] = self.prepare_config_and_inputs() self.check_save_load(**lowercase ) def A_ ( self ): _lowerCamelCase : Any = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**lowercase ) @slow def A_ ( self ): _lowerCamelCase, _lowerCamelCase : int = self.get_pretrained_model_and_inputs() _lowerCamelCase : str = model_a(**lowercase ) _lowerCamelCase : List[Any] = outputs[0].numpy() with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(lowercase ) _lowerCamelCase : List[Any] = TFVisionTextDualEncoderModel.from_pretrained(lowercase ) _lowerCamelCase : Union[str, Any] = model_a(**lowercase ) _lowerCamelCase : Optional[Any] = after_outputs[0].numpy() _lowerCamelCase : Tuple = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowercase , 1E-5 ) @require_tf class lowerCAmelCase__ ( lowercase, unittest.TestCase ): '''simple docstring''' def A_ ( self ): _lowerCamelCase : Optional[Any] = TFVisionTextDualEncoderModel.from_vision_text_pretrained( 'hf-internal-testing/tiny-random-vit' , 'hf-internal-testing/tiny-random-bert' ) _lowerCamelCase : Tuple = 13 _lowerCamelCase : Tuple = floats_tensor( [ batch_size, model.vision_model.config.num_channels, model.vision_model.config.image_size, model.vision_model.config.image_size, ] ) _lowerCamelCase : Dict = ids_tensor([batch_size, 4] , model.text_model.config.vocab_size ) _lowerCamelCase : int = random_attention_mask([batch_size, 4] ) _lowerCamelCase : Any = {'pixel_values': pixel_values, 'input_ids': input_ids, 'attention_mask': attention_mask} return model, inputs def A_ ( self , lowercase , lowercase ): _lowerCamelCase : int = TFViTModel(lowercase , name='vision_model' ) _lowerCamelCase : Union[str, Any] = TFBertModel(lowercase , name='text_model' ) return vision_model, text_model def A_ ( self ): _lowerCamelCase : Optional[int] = TFViTModelTester(self ) _lowerCamelCase : str = TFBertModelTester(self ) _lowerCamelCase : int = vit_model_tester.prepare_config_and_inputs() _lowerCamelCase : str = bert_model_tester.prepare_config_and_inputs() _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : Dict = vision_config_and_inputs ( ( _lowerCamelCase ), ( _lowerCamelCase ), ( _lowerCamelCase ), ( _lowerCamelCase ), ( _lowerCamelCase ), ( _lowerCamelCase ), ( _lowerCamelCase ), ) : Union[str, Any] = text_config_and_inputs return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": input_mask, "input_ids": input_ids, "text_token_type_ids": token_type_ids, "text_sequence_labels": sequence_labels, "text_token_labels": token_labels, "text_choice_labels": choice_labels, } @require_tf class lowerCAmelCase__ ( lowercase, unittest.TestCase ): '''simple docstring''' def A_ ( self ): # DeiT repo doesn't have TF weights, but we don't actually use the weights at all so let's # just reinitialize it. _lowerCamelCase : int = TFVisionTextDualEncoderModel.from_vision_text_pretrained( 'Rocketknight1/tiny-random-deit-tf' , 'hf-internal-testing/tiny-random-roberta' ) _lowerCamelCase : Tuple = 13 _lowerCamelCase : Tuple = floats_tensor( [ batch_size, model.vision_model.config.num_channels, model.vision_model.config.image_size, model.vision_model.config.image_size, ] ) _lowerCamelCase : Tuple = ids_tensor([batch_size, 4] , model.text_model.config.vocab_size ) _lowerCamelCase : str = random_attention_mask([batch_size, 4] ) _lowerCamelCase : Optional[Any] = {'pixel_values': pixel_values, 'input_ids': input_ids, 'attention_mask': attention_mask} return model, inputs def A_ ( self , lowercase , lowercase , lowercase , lowercase , lowercase=None , **lowercase ): _lowerCamelCase, _lowerCamelCase : Union[str, Any] = self.get_vision_text_model(lowercase , lowercase ) _lowerCamelCase : Dict = TFVisionTextDualEncoderModel(vision_model=lowercase , text_model=lowercase ) _lowerCamelCase : List[Any] = model( input_ids=lowercase , pixel_values=lowercase , attention_mask=lowercase , output_attentions=lowercase ) _lowerCamelCase : Union[str, Any] = output.vision_model_output.attentions self.assertEqual(len(lowercase ) , vision_config.num_hidden_layers ) # in DEiT, the seq_len equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) _lowerCamelCase : Any = to_atuple(vision_model.config.image_size ) _lowerCamelCase : Union[str, Any] = to_atuple(vision_model.config.patch_size ) _lowerCamelCase : Union[str, Any] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) _lowerCamelCase : Optional[int] = num_patches + 2 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len) ) _lowerCamelCase : List[Any] = output.text_model_output.attentions self.assertEqual(len(lowercase ) , text_config.num_hidden_layers ) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def A_ ( self , lowercase , lowercase ): _lowerCamelCase : List[Any] = TFDeiTModel(lowercase , name='vision_model' ) _lowerCamelCase : List[str] = TFRobertaModel(lowercase , name='text_model' ) return vision_model, text_model def A_ ( self ): _lowerCamelCase : Any = TFDeiTModelTester(self ) _lowerCamelCase : Union[str, Any] = TFRobertaModelTester(self ) _lowerCamelCase : Optional[int] = vit_model_tester.prepare_config_and_inputs() _lowerCamelCase : int = bert_model_tester.prepare_config_and_inputs() _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : Dict = vision_config_and_inputs ( ( _lowerCamelCase ), ( _lowerCamelCase ), ( _lowerCamelCase ), ( _lowerCamelCase ), ( _lowerCamelCase ), ( _lowerCamelCase ), ( _lowerCamelCase ), ) : Dict = text_config_and_inputs return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": input_mask, "input_ids": input_ids, "text_token_type_ids": token_type_ids, "text_sequence_labels": sequence_labels, "text_token_labels": token_labels, "text_choice_labels": choice_labels, } @require_tf class lowerCAmelCase__ ( lowercase, unittest.TestCase ): '''simple docstring''' def A_ ( self ): _lowerCamelCase : Dict = TFVisionTextDualEncoderModel.from_vision_text_pretrained( 'Rocketknight1/tiny-random-clip-tf' , 'hf-internal-testing/tiny-random-bert' ) _lowerCamelCase : Any = 13 _lowerCamelCase : List[Any] = floats_tensor( [ batch_size, model.vision_model.config.num_channels, model.vision_model.config.image_size, model.vision_model.config.image_size, ] ) _lowerCamelCase : Any = ids_tensor([batch_size, 4] , model.text_model.config.vocab_size ) _lowerCamelCase : List[Any] = random_attention_mask([batch_size, 4] ) _lowerCamelCase : List[Any] = {'pixel_values': pixel_values, 'input_ids': input_ids, 'attention_mask': attention_mask} return model, inputs def A_ ( self , lowercase , lowercase ): _lowerCamelCase : Optional[int] = TFCLIPVisionModel(lowercase , name='vision_model' ) _lowerCamelCase : List[str] = TFBertModel(lowercase , name='text_model' ) return vision_model, text_model def A_ ( self ): _lowerCamelCase : Union[str, Any] = TFCLIPVisionModelTester(self ) _lowerCamelCase : Optional[Any] = TFBertModelTester(self ) _lowerCamelCase : str = clip_model_tester.prepare_config_and_inputs() _lowerCamelCase : Optional[int] = bert_model_tester.prepare_config_and_inputs() _lowerCamelCase, _lowerCamelCase : List[Any] = vision_config_and_inputs ( ( _lowerCamelCase ), ( _lowerCamelCase ), ( _lowerCamelCase ), ( _lowerCamelCase ), ( _lowerCamelCase ), ( _lowerCamelCase ), ( _lowerCamelCase ), ) : List[str] = text_config_and_inputs return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": input_mask, "input_ids": input_ids, "text_token_type_ids": token_type_ids, "text_sequence_labels": sequence_labels, "text_token_labels": token_labels, "text_choice_labels": choice_labels, } @require_vision @require_tf class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' @slow def A_ ( self ): _lowerCamelCase : List[str] = TFVisionTextDualEncoderModel.from_pretrained( 'clip-italian/clip-italian' , logit_scale_init_value=1.0 , from_pt=lowercase ) _lowerCamelCase : str = VisionTextDualEncoderProcessor.from_pretrained('clip-italian/clip-italian' ) _lowerCamelCase : Optional[Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) _lowerCamelCase : List[Any] = processor( text=['una foto di un gatto', 'una foto di un cane'] , images=lowercase , padding=lowercase , return_tensors='np' ) _lowerCamelCase : List[Any] = model(**lowercase ) # verify the logits self.assertEqual(outputs.logits_per_image.shape , (inputs.pixel_values.shape[0], inputs.input_ids.shape[0]) ) self.assertEqual( outputs.logits_per_text.shape , (inputs.input_ids.shape[0], inputs.pixel_values.shape[0]) , ) _lowerCamelCase : int = np.array([[1.2_28_47_27, 0.3_10_41_22]] ) self.assertTrue(np.allclose(outputs.logits_per_image.numpy() , lowercase , atol=1E-3 ) )
96
"""simple docstring""" # Imports import numpy as np class lowerCAmelCase__ : '''simple docstring''' def __init__( self , lowercase=None , lowercase=None , lowercase=None , lowercase=None , lowercase=None ): self.set_matricies(red=lowercase , green=lowercase , blue=lowercase , red_edge=lowercase , nir=lowercase ) def A_ ( self , lowercase=None , lowercase=None , lowercase=None , lowercase=None , lowercase=None ): if red is not None: _lowerCamelCase : Optional[int] = red if green is not None: _lowerCamelCase : Optional[Any] = green if blue is not None: _lowerCamelCase : Tuple = blue if red_edge is not None: _lowerCamelCase : Optional[Any] = red_edge if nir is not None: _lowerCamelCase : Union[str, Any] = nir return True def A_ ( self , lowercase="" , lowercase=None , lowercase=None , lowercase=None , lowercase=None , lowercase=None ): self.set_matricies(red=lowercase , green=lowercase , blue=lowercase , red_edge=lowercase , nir=lowercase ) _lowerCamelCase : str = { 'ARVI2': self.arvaa, 'CCCI': self.ccci, 'CVI': self.cvi, 'GLI': self.gli, 'NDVI': self.ndvi, 'BNDVI': self.bndvi, 'redEdgeNDVI': self.red_edge_ndvi, 'GNDVI': self.gndvi, 'GBNDVI': self.gbndvi, 'GRNDVI': self.grndvi, 'RBNDVI': self.rbndvi, 'PNDVI': self.pndvi, 'ATSAVI': self.atsavi, 'BWDRVI': self.bwdrvi, 'CIgreen': self.ci_green, 'CIrededge': self.ci_rededge, 'CI': self.ci, 'CTVI': self.ctvi, 'GDVI': self.gdvi, 'EVI': self.evi, 'GEMI': self.gemi, 'GOSAVI': self.gosavi, 'GSAVI': self.gsavi, 'Hue': self.hue, 'IVI': self.ivi, 'IPVI': self.ipvi, 'I': self.i, 'RVI': self.rvi, 'MRVI': self.mrvi, 'MSAVI': self.m_savi, 'NormG': self.norm_g, 'NormNIR': self.norm_nir, 'NormR': self.norm_r, 'NGRDI': self.ngrdi, 'RI': self.ri, 'S': self.s, 'IF': self._if, 'DVI': self.dvi, 'TVI': self.tvi, 'NDRE': self.ndre, } try: return funcs[index]() except KeyError: print('Index not in the list!' ) return False def A_ ( self ): return -0.18 + (1.17 * ((self.nir - self.red) / (self.nir + self.red))) def A_ ( self ): return ((self.nir - self.redEdge) / (self.nir + self.redEdge)) / ( (self.nir - self.red) / (self.nir + self.red) ) def A_ ( self ): return self.nir * (self.red / (self.green**2)) def A_ ( self ): return (2 * self.green - self.red - self.blue) / ( 2 * self.green + self.red + self.blue ) def A_ ( self ): return (self.nir - self.red) / (self.nir + self.red) def A_ ( self ): return (self.nir - self.blue) / (self.nir + self.blue) def A_ ( self ): return (self.redEdge - self.red) / (self.redEdge + self.red) def A_ ( self ): return (self.nir - self.green) / (self.nir + self.green) def A_ ( self ): return (self.nir - (self.green + self.blue)) / ( self.nir + (self.green + self.blue) ) def A_ ( self ): return (self.nir - (self.green + self.red)) / ( self.nir + (self.green + self.red) ) def A_ ( self ): return (self.nir - (self.blue + self.red)) / (self.nir + (self.blue + self.red)) def A_ ( self ): return (self.nir - (self.green + self.red + self.blue)) / ( self.nir + (self.green + self.red + self.blue) ) def A_ ( self , lowercase=0.08 , lowercase=1.22 , lowercase=0.03 ): return a * ( (self.nir - a * self.red - b) / (a * self.nir + self.red - a * b + x * (1 + a**2)) ) def A_ ( self ): return (0.1 * self.nir - self.blue) / (0.1 * self.nir + self.blue) def A_ ( self ): return (self.nir / self.green) - 1 def A_ ( self ): return (self.nir / self.redEdge) - 1 def A_ ( self ): return (self.red - self.blue) / self.red def A_ ( self ): _lowerCamelCase : Any = self.ndvi() return ((ndvi + 0.5) / (abs(ndvi + 0.5 ))) * (abs(ndvi + 0.5 ) ** (1 / 2)) def A_ ( self ): return self.nir - self.green def A_ ( self ): return 2.5 * ( (self.nir - self.red) / (self.nir + 6 * self.red - 7.5 * self.blue + 1) ) def A_ ( self ): _lowerCamelCase : Any = (2 * (self.nir**2 - self.red**2) + 1.5 * self.nir + 0.5 * self.red) / ( self.nir + self.red + 0.5 ) return n * (1 - 0.25 * n) - (self.red - 0.1_25) / (1 - self.red) def A_ ( self , lowercase=0.16 ): return (self.nir - self.green) / (self.nir + self.green + y) def A_ ( self , lowercase=0.5 ): return ((self.nir - self.green) / (self.nir + self.green + n)) * (1 + n) def A_ ( self ): return np.arctan( ((2 * self.red - self.green - self.blue) / 30.5) * (self.green - self.blue) ) def A_ ( self , lowercase=None , lowercase=None ): return (self.nir - b) / (a * self.red) def A_ ( self ): return (self.nir / ((self.nir + self.red) / 2)) * (self.ndvi() + 1) def A_ ( self ): return (self.red + self.green + self.blue) / 30.5 def A_ ( self ): return self.nir / self.red def A_ ( self ): return (self.rvi() - 1) / (self.rvi() + 1) def A_ ( self ): return ( (2 * self.nir + 1) - ((2 * self.nir + 1) ** 2 - 8 * (self.nir - self.red)) ** (1 / 2) ) / 2 def A_ ( self ): return self.green / (self.nir + self.red + self.green) def A_ ( self ): return self.nir / (self.nir + self.red + self.green) def A_ ( self ): return self.red / (self.nir + self.red + self.green) def A_ ( self ): return (self.green - self.red) / (self.green + self.red) def A_ ( self ): return (self.red - self.green) / (self.red + self.green) def A_ ( self ): _lowerCamelCase : Union[str, Any] = np.max([np.max(self.red ), np.max(self.green ), np.max(self.blue )] ) _lowerCamelCase : Dict = np.min([np.min(self.red ), np.min(self.green ), np.min(self.blue )] ) return (max_value - min_value) / max_value def A_ ( self ): return (2 * self.red - self.green - self.blue) / (self.green - self.blue) def A_ ( self ): return self.nir / self.red def A_ ( self ): return (self.ndvi() + 0.5) ** (1 / 2) def A_ ( self ): return (self.nir - self.redEdge) / (self.nir + self.redEdge)
96
1
"""simple docstring""" import math def _snake_case ( lowercase__ ): _lowerCamelCase : Any = [True] * n _lowerCamelCase : List[Any] = False _lowerCamelCase : Optional[int] = False _lowerCamelCase : Optional[int] = True for i in range(3 , int(n**0.5 + 1 ) , 2 ): _lowerCamelCase : Union[str, Any] = i * 2 while index < n: _lowerCamelCase : List[Any] = False _lowerCamelCase : str = index + i _lowerCamelCase : Any = [2] for i in range(3 , lowercase__ , 2 ): if is_prime[i]: primes.append(lowercase__ ) return primes def _snake_case ( lowercase__ = 999966663333 ): _lowerCamelCase : Tuple = math.floor(math.sqrt(lowercase__ ) ) + 100 _lowerCamelCase : Optional[int] = prime_sieve(lowercase__ ) _lowerCamelCase : List[Any] = 0 _lowerCamelCase : Optional[int] = 0 _lowerCamelCase : Tuple = primes[prime_index] while (last_prime**2) <= limit: _lowerCamelCase : List[str] = primes[prime_index + 1] _lowerCamelCase : Dict = last_prime**2 _lowerCamelCase : int = next_prime**2 # Get numbers divisible by lps(current) _lowerCamelCase : Any = lower_bound + last_prime while upper_bound > current <= limit: matches_sum += current current += last_prime # Reset the upper_bound while (upper_bound - next_prime) > limit: upper_bound -= next_prime # Add the numbers divisible by ups(current) _lowerCamelCase : str = upper_bound - next_prime while current > lower_bound: matches_sum += current current -= next_prime # Remove the numbers divisible by both ups and lps _lowerCamelCase : int = 0 while upper_bound > current <= limit: if current <= lower_bound: # Increment the current number current += last_prime * next_prime continue if current > limit: break # Remove twice since it was added by both ups and lps matches_sum -= current * 2 # Increment the current number current += last_prime * next_prime # Setup for next pair _lowerCamelCase : str = next_prime prime_index += 1 return matches_sum if __name__ == "__main__": print(solution())
96
"""simple docstring""" import torch from torch import nn from transformers import CLIPPreTrainedModel, CLIPVisionModel from ...models.attention import BasicTransformerBlock from ...utils import logging lowercase__ = logging.get_logger(__name__) # pylint: disable=invalid-name class lowerCAmelCase__ ( lowercase ): '''simple docstring''' def __init__( self , lowercase , lowercase=768 ): super().__init__(lowercase ) _lowerCamelCase : Any = proj_size _lowerCamelCase : Dict = CLIPVisionModel(lowercase ) _lowerCamelCase : List[str] = PaintByExampleMapper(lowercase ) _lowerCamelCase : Optional[Any] = nn.LayerNorm(config.hidden_size ) _lowerCamelCase : int = nn.Linear(config.hidden_size , self.proj_size ) # uncondition for scaling _lowerCamelCase : str = nn.Parameter(torch.randn((1, 1, self.proj_size) ) ) def A_ ( self , lowercase , lowercase=False ): _lowerCamelCase : Union[str, Any] = self.model(pixel_values=lowercase ) _lowerCamelCase : int = clip_output.pooler_output _lowerCamelCase : str = self.mapper(latent_states[:, None] ) _lowerCamelCase : List[Any] = self.final_layer_norm(lowercase ) _lowerCamelCase : Dict = self.proj_out(lowercase ) if return_uncond_vector: return latent_states, self.uncond_vector return latent_states class lowerCAmelCase__ ( nn.Module ): '''simple docstring''' def __init__( self , lowercase ): super().__init__() _lowerCamelCase : Tuple = (config.num_hidden_layers + 1) // 5 _lowerCamelCase : int = config.hidden_size _lowerCamelCase : Optional[Any] = 1 _lowerCamelCase : str = nn.ModuleList( [ BasicTransformerBlock(lowercase , lowercase , lowercase , activation_fn='gelu' , attention_bias=lowercase ) for _ in range(lowercase ) ] ) def A_ ( self , lowercase ): for block in self.blocks: _lowerCamelCase : Tuple = block(lowercase ) return hidden_states
96
1
"""simple docstring""" import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py lowercase__ = """.""" # Internal TensorFlow ops that can be safely ignored (mostly specific to a saved model) lowercase__ = [ """Assert""", """AssignVariableOp""", """EmptyTensorList""", """MergeV2Checkpoints""", """ReadVariableOp""", """ResourceGather""", """RestoreV2""", """SaveV2""", """ShardedFilename""", """StatefulPartitionedCall""", """StaticRegexFullMatch""", """VarHandleOp""", ] def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : Dict = SavedModel() _lowerCamelCase : Optional[int] = [] with open(os.path.join(lowercase__ , 'utils' , 'tf_ops' , 'onnx.json' ) ) as f: _lowerCamelCase : Any = json.load(lowercase__ )['opsets'] for i in range(1 , opset + 1 ): onnx_ops.extend(onnx_opsets[str(lowercase__ )] ) with open(lowercase__ , 'rb' ) as f: saved_model.ParseFromString(f.read() ) _lowerCamelCase : List[str] = set() # Iterate over every metagraph in case there is more than one (a saved model can contain multiple graphs) for meta_graph in saved_model.meta_graphs: # Add operations in the graph definition model_op_names.update(node.op for node in meta_graph.graph_def.node ) # Go through the functions in the graph definition for func in meta_graph.graph_def.library.function: # Add operations in each function model_op_names.update(node.op for node in func.node_def ) # Convert to list, sorted if you want _lowerCamelCase : Union[str, Any] = sorted(lowercase__ ) _lowerCamelCase : Optional[int] = [] for op in model_op_names: if op not in onnx_ops and op not in INTERNAL_OPS: incompatible_ops.append(lowercase__ ) if strict and len(lowercase__ ) > 0: raise Exception(f'''Found the following incompatible ops for the opset {opset}:\n''' + incompatible_ops ) elif len(lowercase__ ) > 0: print(f'''Found the following incompatible ops for the opset {opset}:''' ) print(*lowercase__ , sep='\n' ) else: print(f'''The saved model {saved_model_path} can properly be converted with ONNX.''' ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() parser.add_argument("""--saved_model_path""", help="""Path of the saved model to check (the .pb file).""") parser.add_argument( """--opset""", default=12, type=int, help="""The ONNX opset against which the model has to be tested.""" ) parser.add_argument( """--framework""", choices=["""onnx"""], default="""onnx""", help="""Frameworks against which to test the saved model.""" ) parser.add_argument( """--strict""", action="""store_true""", help="""Whether make the checking strict (raise errors) or not (raise warnings)""" ) lowercase__ = parser.parse_args() if args.framework == "onnx": onnx_compliancy(args.saved_model_path, args.strict, args.opset)
96
"""simple docstring""" lowercase__ = { """meter""": """m""", """kilometer""": """km""", """megametre""": """Mm""", """gigametre""": """Gm""", """terametre""": """Tm""", """petametre""": """Pm""", """exametre""": """Em""", """zettametre""": """Zm""", """yottametre""": """Ym""", } # Exponent of the factor(meter) lowercase__ = { """m""": 0, """km""": 3, """Mm""": 6, """Gm""": 9, """Tm""": 12, """Pm""": 15, """Em""": 18, """Zm""": 21, """Ym""": 24, } def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : List[Any] = from_type.lower().strip('s' ) _lowerCamelCase : List[Any] = to_type.lower().strip('s' ) _lowerCamelCase : Optional[int] = UNIT_SYMBOL.get(lowercase__ , lowercase__ ) _lowerCamelCase : Any = UNIT_SYMBOL.get(lowercase__ , lowercase__ ) if from_sanitized not in METRIC_CONVERSION: _lowerCamelCase : Tuple = ( f'''Invalid \'from_type\' value: {from_type!r}.\n''' f'''Conversion abbreviations are: {', '.join(lowercase__ )}''' ) raise ValueError(lowercase__ ) if to_sanitized not in METRIC_CONVERSION: _lowerCamelCase : Any = ( f'''Invalid \'to_type\' value: {to_type!r}.\n''' f'''Conversion abbreviations are: {', '.join(lowercase__ )}''' ) raise ValueError(lowercase__ ) _lowerCamelCase : List[Any] = METRIC_CONVERSION[from_sanitized] _lowerCamelCase : int = METRIC_CONVERSION[to_sanitized] _lowerCamelCase : List[str] = 1 if from_exponent > to_exponent: _lowerCamelCase : List[str] = from_exponent - to_exponent else: _lowerCamelCase : List[Any] = -(to_exponent - from_exponent) return value * pow(10 , lowercase__ ) if __name__ == "__main__": from doctest import testmod testmod()
96
1
"""simple docstring""" import numpy as np lowercase__ = [ ["""a""", """b""", """c""", """d""", """e"""], ["""f""", """g""", """h""", """i""", """k"""], ["""l""", """m""", """n""", """o""", """p"""], ["""q""", """r""", """s""", """t""", """u"""], ["""v""", """w""", """x""", """y""", """z"""], ] class lowerCAmelCase__ : '''simple docstring''' def __init__( self ): _lowerCamelCase : Dict = np.array(lowercase ) def A_ ( self , lowercase ): _lowerCamelCase, _lowerCamelCase : Any = np.where(letter == self.SQUARE ) _lowerCamelCase : Tuple = np.concatenate([indexa + 1, indexa + 1] ) return indexes def A_ ( self , lowercase , lowercase ): _lowerCamelCase : Tuple = self.SQUARE[indexa - 1, indexa - 1] return letter def A_ ( self , lowercase ): _lowerCamelCase : int = message.lower() _lowerCamelCase : Optional[int] = message.replace(' ' , '' ) _lowerCamelCase : List[str] = message.replace('j' , 'i' ) _lowerCamelCase : Any = np.empty((2, len(lowercase )) ) for letter_index in range(len(lowercase ) ): _lowerCamelCase : Dict = self.letter_to_numbers(message[letter_index] ) _lowerCamelCase : str = numbers[0] _lowerCamelCase : Union[str, Any] = numbers[1] _lowerCamelCase : str = first_step.reshape(2 * len(lowercase ) ) _lowerCamelCase : Dict = '' for numbers_index in range(len(lowercase ) ): _lowerCamelCase : str = int(second_step[numbers_index * 2] ) _lowerCamelCase : List[Any] = int(second_step[(numbers_index * 2) + 1] ) _lowerCamelCase : List[Any] = self.numbers_to_letter(lowercase , lowercase ) _lowerCamelCase : Any = encoded_message + letter return encoded_message def A_ ( self , lowercase ): _lowerCamelCase : Tuple = message.lower() message.replace(' ' , '' ) _lowerCamelCase : List[str] = np.empty(2 * len(lowercase ) ) for letter_index in range(len(lowercase ) ): _lowerCamelCase : str = self.letter_to_numbers(message[letter_index] ) _lowerCamelCase : Tuple = numbers[0] _lowerCamelCase : int = numbers[1] _lowerCamelCase : Union[str, Any] = first_step.reshape((2, len(lowercase )) ) _lowerCamelCase : List[str] = '' for numbers_index in range(len(lowercase ) ): _lowerCamelCase : Tuple = int(second_step[0, numbers_index] ) _lowerCamelCase : Any = int(second_step[1, numbers_index] ) _lowerCamelCase : Optional[Any] = self.numbers_to_letter(lowercase , lowercase ) _lowerCamelCase : List[str] = decoded_message + letter return decoded_message
96
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ = {"""configuration_vit_msn""": ["""VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ViTMSNConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ """VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST""", """ViTMSNModel""", """ViTMSNForImageClassification""", """ViTMSNPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_vit_msn import VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMSNConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_msn import ( VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMSNForImageClassification, ViTMSNModel, ViTMSNPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
96
1
"""simple docstring""" # Usage: # ./gen-card-allenai-wmt16.py import os from pathlib import Path def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : int = { 'en': 'Machine learning is great, isn\'t it?', 'ru': 'Машинное обучение - это здорово, не так ли?', 'de': 'Maschinelles Lernen ist großartig, nicht wahr?', } # BLUE scores as follows: # "pair": [fairseq, transformers] _lowerCamelCase : List[Any] = { 'wmt16-en-de-dist-12-1': [2_8.3, 2_7.5_2], 'wmt16-en-de-dist-6-1': [2_7.4, 2_7.1_1], 'wmt16-en-de-12-1': [2_6.9, 2_5.7_5], } _lowerCamelCase : str = f'''{src_lang}-{tgt_lang}''' _lowerCamelCase : Tuple = f''' --- language: - {src_lang} - {tgt_lang} thumbnail: tags: - translation - wmt16 - allenai license: apache-2.0 datasets: - wmt16 metrics: - bleu --- # FSMT ## Model description This is a ported version of fairseq-based [wmt16 transformer](https://github.com/jungokasai/deep-shallow/) for {src_lang}-{tgt_lang}. For more details, please, see [Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation](https://arxiv.org/abs/2006.10369). All 3 models are available: * [wmt16-en-de-dist-12-1](https://huggingface.co/allenai/wmt16-en-de-dist-12-1) * [wmt16-en-de-dist-6-1](https://huggingface.co/allenai/wmt16-en-de-dist-6-1) * [wmt16-en-de-12-1](https://huggingface.co/allenai/wmt16-en-de-12-1) ## Intended uses & limitations #### How to use ```python from transformers import FSMTForConditionalGeneration, FSMTTokenizer mname = "allenai/{model_name}" tokenizer = FSMTTokenizer.from_pretrained(mname) model = FSMTForConditionalGeneration.from_pretrained(mname) input = "{texts[src_lang]}" input_ids = tokenizer.encode(input, return_tensors="pt") outputs = model.generate(input_ids) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) print(decoded) # {texts[tgt_lang]} ``` #### Limitations and bias ## Training data Pretrained weights were left identical to the original model released by allenai. For more details, please, see the [paper](https://arxiv.org/abs/2006.10369). ## Eval results Here are the BLEU scores: model | fairseq | transformers -------|---------|---------- {model_name} | {scores[model_name][0]} | {scores[model_name][1]} The score is slightly below the score reported in the paper, as the researchers don\'t use `sacrebleu` and measure the score on tokenized outputs. `transformers` score was measured using `sacrebleu` on detokenized outputs. The score was calculated using this code: ```bash git clone https://github.com/huggingface/transformers cd transformers export PAIR={pair} export DATA_DIR=data/$PAIR export SAVE_DIR=data/$PAIR export BS=8 export NUM_BEAMS=5 mkdir -p $DATA_DIR sacrebleu -t wmt16 -l $PAIR --echo src > $DATA_DIR/val.source sacrebleu -t wmt16 -l $PAIR --echo ref > $DATA_DIR/val.target echo $PAIR PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py allenai/{model_name} $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS ``` ## Data Sources - [training, etc.](http://www.statmt.org/wmt16/) - [test set](http://matrix.statmt.org/test_sets/newstest2016.tgz?1504722372) ### BibTeX entry and citation info ``` @misc{{kasai2020deep, title={{Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation}}, author={{Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith}}, year={{2020}}, eprint={{2006.10369}}, archivePrefix={{arXiv}}, primaryClass={{cs.CL}} }} ``` ''' model_card_dir.mkdir(parents=lowercase__ , exist_ok=lowercase__ ) _lowerCamelCase : int = os.path.join(lowercase__ , 'README.md' ) print(f'''Generating {path}''' ) with open(lowercase__ , 'w' , encoding='utf-8' ) as f: f.write(lowercase__ ) # make sure we are under the root of the project lowercase__ = Path(__file__).resolve().parent.parent.parent lowercase__ = repo_dir / """model_cards""" for model_name in ["wmt16-en-de-dist-12-1", "wmt16-en-de-dist-6-1", "wmt16-en-de-12-1"]: lowercase__ = model_cards_dir / """allenai""" / model_name write_model_card(model_card_dir, src_lang="""en""", tgt_lang="""de""", model_name=model_name)
96
"""simple docstring""" def _snake_case ( lowercase__ , lowercase__ ): # "extended trapezoidal rule" # int(f) = dx/2 * (f1 + 2f2 + ... + fn) _lowerCamelCase : List[Any] = (boundary[1] - boundary[0]) / steps _lowerCamelCase : Tuple = boundary[0] _lowerCamelCase : Dict = boundary[1] _lowerCamelCase : List[Any] = make_points(lowercase__ , lowercase__ , lowercase__ ) _lowerCamelCase : List[Any] = 0.0 y += (h / 2.0) * f(lowercase__ ) for i in x_i: # print(i) y += h * f(lowercase__ ) y += (h / 2.0) * f(lowercase__ ) return y def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : str = a + h while x < (b - h): yield x _lowerCamelCase : int = x + h def _snake_case ( lowercase__ ): # enter your function here _lowerCamelCase : Optional[Any] = (x - 0) * (x - 0) return y def _snake_case ( ): _lowerCamelCase : int = 0.0 # Lower bound of integration _lowerCamelCase : Optional[int] = 1.0 # Upper bound of integration _lowerCamelCase : List[str] = 1_0.0 # define number of steps or resolution _lowerCamelCase : List[Any] = [a, b] # define boundary of integration _lowerCamelCase : Optional[Any] = method_a(lowercase__ , lowercase__ ) print(f'''y = {y}''' ) if __name__ == "__main__": main()
96
1
"""simple docstring""" import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class lowerCAmelCase__ ( lowercase, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = BertJapaneseTokenizer lowerCamelCase__ = False lowerCamelCase__ = True def A_ ( self ): super().setUp() _lowerCamelCase : Optional[Any] = [ '[UNK]', '[CLS]', '[SEP]', 'こんにちは', 'こん', 'にちは', 'ばんは', '##こん', '##にちは', '##ばんは', '世界', '##世界', '、', '##、', '。', '##。', ] _lowerCamelCase : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def A_ ( self , lowercase ): _lowerCamelCase : int = 'こんにちは、世界。 \nこんばんは、世界。' _lowerCamelCase : Union[str, Any] = 'こんにちは 、 世界 。 こんばんは 、 世界 。' return input_text, output_text def A_ ( self , lowercase ): _lowerCamelCase, _lowerCamelCase : Any = self.get_input_output_texts(lowercase ) _lowerCamelCase : Tuple = tokenizer.encode(lowercase , add_special_tokens=lowercase ) _lowerCamelCase : Dict = tokenizer.decode(lowercase , clean_up_tokenization_spaces=lowercase ) return text, ids def A_ ( self ): pass # TODO add if relevant def A_ ( self ): pass # TODO add if relevant def A_ ( self ): pass # TODO add if relevant def A_ ( self ): _lowerCamelCase : Optional[int] = self.tokenizer_class(self.vocab_file ) _lowerCamelCase : Union[str, Any] = tokenizer.tokenize('こんにちは、世界。\nこんばんは、世界。' ) self.assertListEqual(lowercase , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) def A_ ( self ): _lowerCamelCase : Tuple = self.tokenizer_class(self.vocab_file , word_tokenizer_type='mecab' ) self.assertIsNotNone(lowercase ) _lowerCamelCase : int = 'こんにちは、世界。\nこんばんは、世界。' _lowerCamelCase : Tuple = tokenizer.tokenize(lowercase ) self.assertListEqual(lowercase , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) _lowerCamelCase : Union[str, Any] = os.path.join(self.tmpdirname , 'tokenizer.bin' ) with open(lowercase , 'wb' ) as handle: pickle.dump(lowercase , lowercase ) with open(lowercase , 'rb' ) as handle: _lowerCamelCase : Dict = pickle.load(lowercase ) _lowerCamelCase : Dict = tokenizer_new.tokenize(lowercase ) self.assertListEqual(lowercase , lowercase ) def A_ ( self ): _lowerCamelCase : Union[str, Any] = MecabTokenizer(mecab_dic='ipadic' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def A_ ( self ): try: _lowerCamelCase : Dict = MecabTokenizer(mecab_dic='unidic_lite' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def A_ ( self ): try: _lowerCamelCase : Any = MecabTokenizer(mecab_dic='unidic' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def A_ ( self ): _lowerCamelCase : List[Any] = MecabTokenizer(do_lower_case=lowercase , mecab_dic='ipadic' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iphone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) def A_ ( self ): try: _lowerCamelCase : Tuple = MecabTokenizer( do_lower_case=lowercase , normalize_text=lowercase , mecab_option='-d /usr/local/lib/mecab/dic/jumandic' ) except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れた', '\u3000', '。'] , ) def A_ ( self ): _lowerCamelCase : Union[str, Any] = MecabTokenizer(normalize_text=lowercase , mecab_dic='ipadic' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップルストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', ' ', '。'] , ) @require_sudachi def A_ ( self ): _lowerCamelCase : Optional[int] = self.tokenizer_class(self.vocab_file , word_tokenizer_type='sudachi' ) self.assertIsNotNone(lowercase ) _lowerCamelCase : int = 'こんにちは、世界。\nこんばんは、世界。' _lowerCamelCase : Tuple = tokenizer.tokenize(lowercase ) self.assertListEqual(lowercase , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) _lowerCamelCase : Union[str, Any] = os.path.join(self.tmpdirname , 'tokenizer.bin' ) with open(lowercase , 'wb' ) as handle: pickle.dump(lowercase , lowercase ) with open(lowercase , 'rb' ) as handle: _lowerCamelCase : int = pickle.load(lowercase ) _lowerCamelCase : List[Any] = tokenizer_new.tokenize(lowercase ) self.assertListEqual(lowercase , lowercase ) @require_sudachi def A_ ( self ): _lowerCamelCase : Optional[Any] = SudachiTokenizer(sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , [' ', '\t', 'アップル', 'ストア', 'で', 'iPhone', '8', ' ', 'が', ' ', ' ', '\n ', '発売', 'さ', 'れ', 'た', ' ', '。', ' ', ' '] , ) @require_sudachi def A_ ( self ): _lowerCamelCase : str = SudachiTokenizer(sudachi_dict_type='core' , sudachi_split_mode='A' ) self.assertListEqual(tokenizer.tokenize('外国人参政権' ) , ['外国', '人', '参政', '権'] ) @require_sudachi def A_ ( self ): _lowerCamelCase : Tuple = SudachiTokenizer(sudachi_dict_type='core' , sudachi_split_mode='B' ) self.assertListEqual(tokenizer.tokenize('外国人参政権' ) , ['外国人', '参政権'] ) @require_sudachi def A_ ( self ): _lowerCamelCase : List[Any] = SudachiTokenizer(sudachi_dict_type='core' , sudachi_split_mode='C' ) self.assertListEqual(tokenizer.tokenize('外国人参政権' ) , ['外国人参政権'] ) @require_sudachi def A_ ( self ): _lowerCamelCase : Tuple = SudachiTokenizer(do_lower_case=lowercase , sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , [' ', '\t', 'アップル', 'ストア', 'で', 'iphone', '8', ' ', 'が', ' ', ' ', '\n ', '発売', 'さ', 'れ', 'た', ' ', '。', ' ', ' '] , ) @require_sudachi def A_ ( self ): _lowerCamelCase : Tuple = SudachiTokenizer(normalize_text=lowercase , sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , [' ', '\t', 'アップル', 'ストア', 'で', 'iPhone', '8', ' ', 'が', ' ', ' ', '\n ', '発売', 'さ', 'れ', 'た', '\u3000', '。', ' ', ' '] , ) @require_sudachi def A_ ( self ): _lowerCamelCase : str = SudachiTokenizer(trim_whitespace=lowercase , sudachi_dict_type='core' ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れ', 'た', '。'] , ) @require_jumanpp def A_ ( self ): _lowerCamelCase : Union[str, Any] = self.tokenizer_class(self.vocab_file , word_tokenizer_type='jumanpp' ) self.assertIsNotNone(lowercase ) _lowerCamelCase : Optional[Any] = 'こんにちは、世界。\nこんばんは、世界。' _lowerCamelCase : Optional[Any] = tokenizer.tokenize(lowercase ) self.assertListEqual(lowercase , ['こんにちは', '、', '世界', '。', 'こん', '##ばんは', '、', '世界', '。'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) _lowerCamelCase : int = os.path.join(self.tmpdirname , 'tokenizer.bin' ) with open(lowercase , 'wb' ) as handle: pickle.dump(lowercase , lowercase ) with open(lowercase , 'rb' ) as handle: _lowerCamelCase : Tuple = pickle.load(lowercase ) _lowerCamelCase : Dict = tokenizer_new.tokenize(lowercase ) self.assertListEqual(lowercase , lowercase ) @require_jumanpp def A_ ( self ): _lowerCamelCase : List[Any] = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', '\u3000', 'が', '\u3000', '\u3000', '\u3000', '発売', 'さ', 'れた', '\u3000', '。'] , ) @require_jumanpp def A_ ( self ): _lowerCamelCase : str = JumanppTokenizer(do_lower_case=lowercase ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iphone', '8', '\u3000', 'が', '\u3000', '\u3000', '\u3000', '発売', 'さ', 'れた', '\u3000', '。'] , ) @require_jumanpp def A_ ( self ): _lowerCamelCase : List[Any] = JumanppTokenizer(normalize_text=lowercase ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['ア', 'ッ', 'フ', '゚', 'ル', 'ストア', 'で', 'iPhone', '8', '\u3000', 'が', '\u3000', '\u3000', '\u3000', '発売', 'さ', 'れた', '\u3000', '。'] , ) @require_jumanpp def A_ ( self ): _lowerCamelCase : Union[str, Any] = JumanppTokenizer(trim_whitespace=lowercase ) self.assertListEqual( tokenizer.tokenize(' \tアップルストアでiPhone8 が \n 発売された 。 ' ) , ['アップル', 'ストア', 'で', 'iPhone', '8', 'が', '発売', 'さ', 'れた', '。'] , ) @require_jumanpp def A_ ( self ): _lowerCamelCase : Optional[int] = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize('ありがとうございますm(_ _)m見つけるのが大変です。' ) , ['ありがとう', 'ございます', 'm(_ _)m', '見つける', 'の', 'が', '大変です', '。'] , ) def A_ ( self ): _lowerCamelCase : Any = ['[UNK]', '[CLS]', '[SEP]', 'こんにちは', 'こん', 'にちは', 'ばんは', '##こん', '##にちは', '##ばんは'] _lowerCamelCase : int = {} for i, token in enumerate(lowercase ): _lowerCamelCase : List[str] = i _lowerCamelCase : Tuple = WordpieceTokenizer(vocab=lowercase , unk_token='[UNK]' ) self.assertListEqual(tokenizer.tokenize('' ) , [] ) self.assertListEqual(tokenizer.tokenize('こんにちは' ) , ['こんにちは'] ) self.assertListEqual(tokenizer.tokenize('こんばんは' ) , ['こん', '##ばんは'] ) self.assertListEqual(tokenizer.tokenize('こんばんは こんばんにちは こんにちは' ) , ['こん', '##ばんは', '[UNK]', 'こんにちは'] ) def A_ ( self ): _lowerCamelCase : Optional[Any] = BertJapaneseTokenizer.from_pretrained('nlp-waseda/roberta-base-japanese-with-auto-jumanpp' ) _lowerCamelCase : Optional[int] = tokenizer.subword_tokenizer _lowerCamelCase : Union[str, Any] = subword_tokenizer.tokenize('国境 の 長い トンネル を 抜ける と 雪国 であった 。' ) self.assertListEqual(lowercase , ['▁国境', '▁の', '▁長い', '▁トンネル', '▁を', '▁抜ける', '▁と', '▁雪', '国', '▁であった', '▁。'] ) _lowerCamelCase : Optional[int] = subword_tokenizer.tokenize('こんばんは こんばん にち は こんにちは' ) self.assertListEqual(lowercase , ['▁こん', 'ばん', 'は', '▁こん', 'ばん', '▁に', 'ち', '▁は', '▁こんにちは'] ) def A_ ( self ): _lowerCamelCase : str = self.tokenizer_class.from_pretrained('cl-tohoku/bert-base-japanese' ) _lowerCamelCase : int = tokenizer.encode('ありがとう。' , add_special_tokens=lowercase ) _lowerCamelCase : Any = tokenizer.encode('どういたしまして。' , add_special_tokens=lowercase ) _lowerCamelCase : Union[str, Any] = tokenizer.build_inputs_with_special_tokens(lowercase ) _lowerCamelCase : Union[str, Any] = tokenizer.build_inputs_with_special_tokens(lowercase , lowercase ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class lowerCAmelCase__ ( lowercase, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = BertJapaneseTokenizer lowerCamelCase__ = False def A_ ( self ): super().setUp() _lowerCamelCase : int = ['[UNK]', '[CLS]', '[SEP]', 'こ', 'ん', 'に', 'ち', 'は', 'ば', '世', '界', '、', '。'] _lowerCamelCase : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def A_ ( self , **lowercase ): return BertJapaneseTokenizer.from_pretrained(self.tmpdirname , subword_tokenizer_type='character' , **lowercase ) def A_ ( self , lowercase ): _lowerCamelCase : Tuple = 'こんにちは、世界。 \nこんばんは、世界。' _lowerCamelCase : Dict = 'こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。' return input_text, output_text def A_ ( self ): pass # TODO add if relevant def A_ ( self ): pass # TODO add if relevant def A_ ( self ): pass # TODO add if relevant def A_ ( self ): _lowerCamelCase : Dict = self.tokenizer_class(self.vocab_file , subword_tokenizer_type='character' ) _lowerCamelCase : List[Any] = tokenizer.tokenize('こんにちは、世界。 \nこんばんは、世界。' ) self.assertListEqual( lowercase , ['こ', 'ん', 'に', 'ち', 'は', '、', '世', '界', '。', 'こ', 'ん', 'ば', 'ん', 'は', '、', '世', '界', '。'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowercase ) , [3, 4, 5, 6, 7, 11, 9, 10, 12, 3, 4, 8, 4, 7, 11, 9, 10, 12] ) def A_ ( self ): _lowerCamelCase : Union[str, Any] = ['[UNK]', '[CLS]', '[SEP]', 'こ', 'ん', 'に', 'ち', 'は', 'ば', '世', '界', '、', '。'] _lowerCamelCase : Optional[int] = {} for i, token in enumerate(lowercase ): _lowerCamelCase : str = i _lowerCamelCase : List[Any] = CharacterTokenizer(vocab=lowercase , unk_token='[UNK]' ) self.assertListEqual(tokenizer.tokenize('' ) , [] ) self.assertListEqual(tokenizer.tokenize('こんにちは' ) , ['こ', 'ん', 'に', 'ち', 'は'] ) self.assertListEqual(tokenizer.tokenize('こんにちほ' ) , ['こ', 'ん', 'に', 'ち', '[UNK]'] ) def A_ ( self ): _lowerCamelCase : Optional[Any] = self.tokenizer_class.from_pretrained('cl-tohoku/bert-base-japanese-char' ) _lowerCamelCase : str = tokenizer.encode('ありがとう。' , add_special_tokens=lowercase ) _lowerCamelCase : Optional[Any] = tokenizer.encode('どういたしまして。' , add_special_tokens=lowercase ) _lowerCamelCase : Dict = tokenizer.build_inputs_with_special_tokens(lowercase ) _lowerCamelCase : Optional[int] = tokenizer.build_inputs_with_special_tokens(lowercase , lowercase ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def A_ ( self ): _lowerCamelCase : Optional[int] = 'cl-tohoku/bert-base-japanese' _lowerCamelCase : str = AutoTokenizer.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def A_ ( self ): _lowerCamelCase : Any = 'cl-tohoku/bert-base-japanese' with self.assertLogs('transformers' , level='WARNING' ) as cm: BertTokenizer.from_pretrained(lowercase ) self.assertTrue( cm.records[0].message.startswith( 'The tokenizer class you load from this checkpoint is not the same type as the class this function' ' is called from.' ) ) _lowerCamelCase : Optional[Any] = 'bert-base-cased' with self.assertLogs('transformers' , level='WARNING' ) as cm: BertJapaneseTokenizer.from_pretrained(lowercase ) self.assertTrue( cm.records[0].message.startswith( 'The tokenizer class you load from this checkpoint is not the same type as the class this function' ' is called from.' ) )
96
"""simple docstring""" import math def _snake_case ( lowercase__ ): return math.sqrt(lowercase__ ) * math.sqrt(lowercase__ ) == num def _snake_case ( lowercase__ ): _lowerCamelCase : Optional[int] = 0 _lowerCamelCase : List[Any] = n while left <= right: _lowerCamelCase : str = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: _lowerCamelCase : str = mid - 1 else: _lowerCamelCase : Optional[int] = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : Tuple = f'''{sampling_rate}''' _lowerCamelCase : str = '1' _lowerCamelCase : str = 'f32le' _lowerCamelCase : Union[str, Any] = [ 'ffmpeg', '-i', 'pipe:0', '-ac', ac, '-ar', ar, '-f', format_for_conversion, '-hide_banner', '-loglevel', 'quiet', 'pipe:1', ] try: with subprocess.Popen(lowercase__ , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: _lowerCamelCase : str = ffmpeg_process.communicate(lowercase__ ) except FileNotFoundError as error: raise ValueError('ffmpeg was not found but is required to load audio files from filename' ) from error _lowerCamelCase : List[Any] = output_stream[0] _lowerCamelCase : Tuple = np.frombuffer(lowercase__ , np.floataa ) if audio.shape[0] == 0: raise ValueError('Malformed soundfile' ) return audio def _snake_case ( lowercase__ , lowercase__ , lowercase__ = "f32le" , ): _lowerCamelCase : Optional[Any] = f'''{sampling_rate}''' _lowerCamelCase : List[str] = '1' if format_for_conversion == "s16le": _lowerCamelCase : List[str] = 2 elif format_for_conversion == "f32le": _lowerCamelCase : List[Any] = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) _lowerCamelCase : Dict = platform.system() if system == "Linux": _lowerCamelCase : Optional[int] = 'alsa' _lowerCamelCase : Optional[Any] = 'default' elif system == "Darwin": _lowerCamelCase : Optional[int] = 'avfoundation' _lowerCamelCase : Any = ':0' elif system == "Windows": _lowerCamelCase : Tuple = 'dshow' _lowerCamelCase : Tuple = 'default' _lowerCamelCase : Optional[int] = [ 'ffmpeg', '-f', format_, '-i', input_, '-ac', ac, '-ar', ar, '-f', format_for_conversion, '-fflags', 'nobuffer', '-hide_banner', '-loglevel', 'quiet', 'pipe:1', ] _lowerCamelCase : Tuple = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample _lowerCamelCase : List[Any] = _ffmpeg_stream(lowercase__ , lowercase__ ) for item in iterator: yield item def _snake_case ( lowercase__ , lowercase__ , lowercase__ = None , lowercase__ = None , lowercase__ = "f32le" , ): if stream_chunk_s is not None: _lowerCamelCase : int = stream_chunk_s else: _lowerCamelCase : Optional[Any] = chunk_length_s _lowerCamelCase : Optional[Any] = ffmpeg_microphone(lowercase__ , lowercase__ , format_for_conversion=lowercase__ ) if format_for_conversion == "s16le": _lowerCamelCase : List[str] = np.intaa _lowerCamelCase : str = 2 elif format_for_conversion == "f32le": _lowerCamelCase : Any = np.floataa _lowerCamelCase : List[Any] = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) if stride_length_s is None: _lowerCamelCase : Union[str, Any] = chunk_length_s / 6 _lowerCamelCase : Optional[int] = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(lowercase__ , (int, float) ): _lowerCamelCase : Any = [stride_length_s, stride_length_s] _lowerCamelCase : Tuple = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample _lowerCamelCase : Optional[Any] = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample _lowerCamelCase : List[Any] = datetime.datetime.now() _lowerCamelCase : Optional[int] = datetime.timedelta(seconds=lowercase__ ) for item in chunk_bytes_iter(lowercase__ , lowercase__ , stride=(stride_left, stride_right) , stream=lowercase__ ): # Put everything back in numpy scale _lowerCamelCase : List[Any] = np.frombuffer(item['raw'] , dtype=lowercase__ ) _lowerCamelCase : int = ( item['stride'][0] // size_of_sample, item['stride'][1] // size_of_sample, ) _lowerCamelCase : Optional[int] = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ = False ): _lowerCamelCase : int = B'' _lowerCamelCase, _lowerCamelCase : Dict = stride if stride_left + stride_right >= chunk_len: raise ValueError( f'''Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}''' ) _lowerCamelCase : str = 0 for raw in iterator: acc += raw if stream and len(lowercase__ ) < chunk_len: _lowerCamelCase : Optional[int] = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(lowercase__ ) >= chunk_len: # We are flushing the accumulator _lowerCamelCase : str = (_stride_left, stride_right) _lowerCamelCase : str = {'raw': acc[:chunk_len], 'stride': stride} if stream: _lowerCamelCase : List[Any] = False yield item _lowerCamelCase : Optional[Any] = stride_left _lowerCamelCase : str = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(lowercase__ ) > stride_left: _lowerCamelCase : Optional[Any] = {'raw': acc, 'stride': (_stride_left, 0)} if stream: _lowerCamelCase : Tuple = False yield item def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : int = 2**24 # 16Mo try: with subprocess.Popen(lowercase__ , stdout=subprocess.PIPE , bufsize=lowercase__ ) as ffmpeg_process: while True: _lowerCamelCase : Optional[Any] = ffmpeg_process.stdout.read(lowercase__ ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError('ffmpeg was not found but is required to stream audio files from filename' ) from error
96
"""simple docstring""" import functools from typing import Any def _snake_case ( lowercase__ , lowercase__ ): # Validation if not isinstance(lowercase__ , lowercase__ ) or len(lowercase__ ) == 0: raise ValueError('the string should be not empty string' ) if not isinstance(lowercase__ , lowercase__ ) or not all( isinstance(lowercase__ , lowercase__ ) and len(lowercase__ ) > 0 for item in words ): raise ValueError('the words should be a list of non-empty strings' ) # Build trie _lowerCamelCase : dict[str, Any] = {} _lowerCamelCase : List[Any] = 'WORD_KEEPER' for word in words: _lowerCamelCase : Dict = trie for c in word: if c not in trie_node: _lowerCamelCase : Any = {} _lowerCamelCase : str = trie_node[c] _lowerCamelCase : Optional[Any] = True _lowerCamelCase : Dict = len(lowercase__ ) # Dynamic programming method @functools.cache def is_breakable(lowercase__ ) -> bool: if index == len_string: return True _lowerCamelCase : List[Any] = trie for i in range(lowercase__ , lowercase__ ): _lowerCamelCase : Any = trie_node.get(string[i] , lowercase__ ) if trie_node is None: return False if trie_node.get(lowercase__ , lowercase__ ) and is_breakable(i + 1 ): return True return False return is_breakable(0 ) if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" import copy import os import cva import numpy as np from matplotlib import pyplot as plt class lowerCAmelCase__ : '''simple docstring''' def __init__( self ): _lowerCamelCase : List[Any] = '' _lowerCamelCase : Dict = '' _lowerCamelCase : Any = [] _lowerCamelCase : Tuple = 0 _lowerCamelCase : Optional[int] = 256 _lowerCamelCase : Dict = 0 _lowerCamelCase : Union[str, Any] = 0 _lowerCamelCase : Tuple = 0 _lowerCamelCase : List[str] = 0 def A_ ( self , lowercase ): _lowerCamelCase : int = cva.imread(lowercase , 0 ) _lowerCamelCase : int = copy.deepcopy(self.img ) _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : Any = plt.hist(self.img.ravel() , 256 , [0, 256] , label='x' ) _lowerCamelCase : Tuple = np.sum(lowercase ) for i in range(len(lowercase ) ): _lowerCamelCase : Optional[int] = x[i] / self.k self.sk += prk _lowerCamelCase : List[str] = (self.L - 1) * self.sk if self.rem != 0: _lowerCamelCase : int = int(last % last ) _lowerCamelCase : Optional[int] = int(last + 1 if self.rem >= 0.5 else last ) self.last_list.append(lowercase ) _lowerCamelCase : int = int(np.ma.count(self.img ) / self.img[1].size ) _lowerCamelCase : Tuple = self.img[1].size for i in range(self.number_of_cols ): for j in range(self.number_of_rows ): _lowerCamelCase : Optional[int] = self.img[j][i] if num != self.last_list[num]: _lowerCamelCase : int = self.last_list[num] cva.imwrite('output_data/output.jpg' , self.img ) def A_ ( self ): plt.hist(self.img.ravel() , 256 , [0, 256] ) def A_ ( self ): cva.imshow('Output-Image' , self.img ) cva.imshow('Input-Image' , self.original_image ) cva.waitKey(5000 ) cva.destroyAllWindows() if __name__ == "__main__": lowercase__ = os.path.join(os.path.basename(__file__), """image_data/input.jpg""") lowercase__ = ConstantStretch() stretcher.stretch(file_path) stretcher.plot_histogram() stretcher.show_image()
96
"""simple docstring""" def _snake_case ( lowercase__ ): if not isinstance(lowercase__ , lowercase__ ): raise ValueError('Input series is not valid, valid series - [2, 4, 6]' ) if len(lowercase__ ) == 0: raise ValueError('Input list must be a non empty list' ) if len(lowercase__ ) == 1: return True _lowerCamelCase : List[Any] = series[1] - series[0] for index in range(len(lowercase__ ) - 1 ): if series[index + 1] - series[index] != common_diff: return False return True def _snake_case ( lowercase__ ): if not isinstance(lowercase__ , lowercase__ ): raise ValueError('Input series is not valid, valid series - [2, 4, 6]' ) if len(lowercase__ ) == 0: raise ValueError('Input list must be a non empty list' ) _lowerCamelCase : Optional[int] = 0 for val in series: answer += val return answer / len(lowercase__ ) if __name__ == "__main__": import doctest doctest.testmod()
96
1
"""simple docstring""" import logging import re import pytorch_quantization import pytorch_quantization.nn as quant_nn import torch from pytorch_quantization import calib from pytorch_quantization.tensor_quant import QuantDescriptor lowercase__ = logging.getLogger(__name__) lowercase__ = 50 # max width of layer names lowercase__ = 70 # max width of quantizer names def _snake_case ( lowercase__ ): _lowerCamelCase : Dict = parser.add_argument_group('quant_trainer arguments' ) group.add_argument('--wprec' , type=lowercase__ , default=8 , help='weight precision' ) group.add_argument('--aprec' , type=lowercase__ , default=8 , help='activation precision' ) group.add_argument('--quant-per-tensor' , action='store_true' , help='per tensor weight scaling' ) group.add_argument('--quant-disable' , action='store_true' , help='disable all quantizers' ) group.add_argument('--quant-disable-embeddings' , action='store_true' , help='disable all embeddings quantizers' ) group.add_argument('--quant-disable-keyword' , type=lowercase__ , nargs='+' , help='disable quantizers by keyword' ) group.add_argument('--quant-disable-layer-module' , type=lowercase__ , help='disable quantizers by keyword under layer.' ) group.add_argument('--quant-enable-layer-module' , type=lowercase__ , help='enable quantizers by keyword under layer' ) group.add_argument('--calibrator' , default='max' , help='which quantization range calibrator to use' ) group.add_argument('--percentile' , default=lowercase__ , type=lowercase__ , help='percentile for PercentileCalibrator' ) group.add_argument('--fuse-qkv' , action='store_true' , help='use the same scale factor for qkv' ) group.add_argument('--clip-gelu' , metavar='N' , type=lowercase__ , help='clip gelu output maximum value to N' ) group.add_argument( '--recalibrate-weights' , action='store_true' , help=( 'recalibrate weight amaxes by taking the max of the weights.' ' amaxes will be computed with the current quantization granularity (axis).' ) , ) def _snake_case ( lowercase__ ): if args.calibrator == "max": _lowerCamelCase : Union[str, Any] = 'max' elif args.calibrator == "percentile": if args.percentile is None: raise ValueError('Specify --percentile when using percentile calibrator' ) _lowerCamelCase : Tuple = 'histogram' elif args.calibrator == "mse": _lowerCamelCase : Optional[Any] = 'histogram' else: raise ValueError(f'''Invalid calibrator {args.calibrator}''' ) _lowerCamelCase : List[Any] = QuantDescriptor(num_bits=args.aprec , calib_method=lowercase__ ) _lowerCamelCase : int = QuantDescriptor(num_bits=args.wprec , axis=(None if args.quant_per_tensor else (0,)) ) quant_nn.QuantLinear.set_default_quant_desc_input(lowercase__ ) quant_nn.QuantLinear.set_default_quant_desc_weight(lowercase__ ) def _snake_case ( lowercase__ , lowercase__ , lowercase__=False , lowercase__=False ): logger.info('Configuring Model for Quantization' ) logger.info(f'''using quantization package {pytorch_quantization.__file__}''' ) if not calib: if args.quant_disable_embeddings: set_quantizer_by_name(lowercase__ , ['embeddings'] , which='weight' , _disabled=lowercase__ ) if args.quant_disable: set_quantizer_by_name(lowercase__ , [''] , _disabled=lowercase__ ) if args.quant_disable_keyword: set_quantizer_by_name(lowercase__ , args.quant_disable_keyword , _disabled=lowercase__ ) if args.quant_disable_layer_module: set_quantizer_by_name(lowercase__ , [r'layer.\d+.' + args.quant_disable_layer_module] , _disabled=lowercase__ ) if args.quant_enable_layer_module: set_quantizer_by_name(lowercase__ , [r'layer.\d+.' + args.quant_enable_layer_module] , _disabled=lowercase__ ) if args.recalibrate_weights: recalibrate_weights(lowercase__ ) if args.fuse_qkv: fuse_qkv(lowercase__ , lowercase__ ) if args.clip_gelu: clip_gelu(lowercase__ , args.clip_gelu ) # if args.local_rank in [-1, 0] and not calib: print_quant_summary(lowercase__ ) def _snake_case ( lowercase__ ): logger.info('Enabling Calibration' ) for name, module in model.named_modules(): if name.endswith('_quantizer' ): if module._calibrator is not None: module.disable_quant() module.enable_calib() else: module.disable() logger.info(f'''{name:80}: {module}''' ) def _snake_case ( lowercase__ , lowercase__ ): logger.info('Loading calibrated amax' ) for name, module in model.named_modules(): if name.endswith('_quantizer' ): if module._calibrator is not None: if isinstance(module._calibrator , calib.MaxCalibrator ): module.load_calib_amax() else: module.load_calib_amax('percentile' , percentile=args.percentile ) module.enable_quant() module.disable_calib() else: module.enable() model.cuda() print_quant_summary(lowercase__ ) def _snake_case ( lowercase__ , lowercase__ ): def fusea(lowercase__ , lowercase__ , lowercase__ ): for mod in [qq, qk, qv]: if not hasattr(lowercase__ , '_amax' ): print(' WARNING: NO AMAX BUFFER' ) return _lowerCamelCase : Dict = qq._amax.detach().item() _lowerCamelCase : Union[str, Any] = qk._amax.detach().item() _lowerCamelCase : int = qv._amax.detach().item() _lowerCamelCase : Optional[int] = max(lowercase__ , lowercase__ , lowercase__ ) qq._amax.fill_(lowercase__ ) qk._amax.fill_(lowercase__ ) qv._amax.fill_(lowercase__ ) logger.info(f''' q={q:5.2f} k={k:5.2f} v={v:5.2f} -> {amax:5.2f}''' ) for name, mod in model.named_modules(): if name.endswith('.attention.self' ): logger.info(f'''FUSE_QKV: {name:{name_width}}''' ) fusea(mod.matmul_q_input_quantizer , mod.matmul_k_input_quantizer , mod.matmul_v_input_quantizer ) if args.quant_per_tensor: fusea(mod.query._weight_quantizer , mod.key._weight_quantizer , mod.value._weight_quantizer ) def _snake_case ( lowercase__ , lowercase__ ): for name, mod in model.named_modules(): if name.endswith('.output.dense' ) and not name.endswith('attention.output.dense' ): _lowerCamelCase : Optional[int] = mod._input_quantizer._amax.data.detach().item() mod._input_quantizer._amax.data.detach().clamp_(max=lowercase__ ) _lowerCamelCase : List[Any] = mod._input_quantizer._amax.data.detach().item() logger.info(f'''CLIP_GELU: {name:{name_width}} amax: {amax_init:5.2f} -> {amax:5.2f}''' ) def _snake_case ( lowercase__ ): for name, mod in model.named_modules(): if hasattr(lowercase__ , '_weight_quantizer' ) and mod._weight_quantizer.axis is not None: _lowerCamelCase : Union[str, Any] = mod.weight.shape[0] _lowerCamelCase : int = mod._weight_quantizer._amax.detach() _lowerCamelCase : str = torch.ones(lowercase__ , dtype=amax.dtype , device=amax.device ) * amax print(f'''expanding {name} {amax} -> {mod._weight_quantizer._amax}''' ) def _snake_case ( lowercase__ ): for name, mod in model.named_modules(): if hasattr(lowercase__ , '_weight_quantizer' ): if not hasattr(mod.weight_quantizer , '_amax' ): print('RECALIB: {name:{name_width}} WARNING: NO AMAX BUFFER' ) continue # determine which axes to reduce across # e.g. a 4D tensor quantized per axis 0 should reduce over (1,2,3) _lowerCamelCase : List[str] = set() if mod._weight_quantizer.axis is None else set(mod._weight_quantizer.axis ) _lowerCamelCase : str = set(range(len(mod.weight.size() ) ) ) - axis_set _lowerCamelCase : int = pytorch_quantization.utils.reduce_amax(mod.weight , axis=lowercase__ , keepdims=lowercase__ ).detach() logger.info(f'''RECALIB: {name:{name_width}} {mod._weight_quantizer._amax.flatten()} -> {amax.flatten()}''' ) _lowerCamelCase : Any = amax def _snake_case ( lowercase__ , lowercase__=25 , lowercase__=180 , lowercase__=None ): if ignore is None: _lowerCamelCase : str = [] elif not isinstance(lowercase__ , lowercase__ ): _lowerCamelCase : int = [ignore] _lowerCamelCase : Any = 0 for name, mod in model.named_modules(): if not hasattr(lowercase__ , 'weight' ): continue _lowerCamelCase : Optional[Any] = max(lowercase__ , len(lowercase__ ) ) for name, mod in model.named_modules(): _lowerCamelCase : Any = getattr(lowercase__ , '_input_quantizer' , lowercase__ ) _lowerCamelCase : Dict = getattr(lowercase__ , '_weight_quantizer' , lowercase__ ) if not hasattr(lowercase__ , 'weight' ): continue if type(lowercase__ ) in ignore: continue if [True for s in ignore if type(lowercase__ ) is str and s in name]: continue _lowerCamelCase : List[str] = f'''Act:{input_q.extra_repr()}''' _lowerCamelCase : Dict = f'''Wgt:{weight_q.extra_repr()}''' _lowerCamelCase : str = f'''{name:{name_width}} {act_str} {wgt_str}''' if len(lowercase__ ) <= line_width: logger.info(lowercase__ ) else: logger.info(f'''{name:{name_width}} {act_str}''' ) logger.info(f'''{' ':{name_width}} {wgt_str}''' ) def _snake_case ( lowercase__ ): _lowerCamelCase : Dict = 0 for name, mod in model.named_modules(): if isinstance(lowercase__ , pytorch_quantization.nn.TensorQuantizer ): print(f'''{name:80} {mod}''' ) count += 1 print(f'''{count} TensorQuantizers found in model''' ) def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): _lowerCamelCase : Optional[int] = getattr(lowercase__ , lowercase__ , lowercase__ ) if quantizer_mod is not None: assert hasattr(lowercase__ , lowercase__ ) setattr(lowercase__ , lowercase__ , lowercase__ ) else: logger.warning(f'''{name} has no {quantizer}''' ) def _snake_case ( lowercase__ , lowercase__ , lowercase__="both" , **lowercase__ ): _lowerCamelCase : str = f'''Warning: changing {which} quantizers of {name:{qname_width}}''' for k, v in kwargs.items(): s += f''' {k}={v}''' if which in ["input", "both"]: set_quantizer(lowercase__ , lowercase__ , '_input_quantizer' , lowercase__ , lowercase__ ) if which in ["weight", "both"]: set_quantizer(lowercase__ , lowercase__ , '_weight_quantizer' , lowercase__ , lowercase__ ) logger.info(lowercase__ ) def _snake_case ( lowercase__ , lowercase__ , **lowercase__ ): for name, mod in model.named_modules(): if hasattr(lowercase__ , '_input_quantizer' ) or hasattr(lowercase__ , '_weight_quantizer' ): for n in names: if re.search(lowercase__ , lowercase__ ): set_quantizers(lowercase__ , lowercase__ , **lowercase__ ) elif name.endswith('_quantizer' ): for n in names: if re.search(lowercase__ , lowercase__ ): _lowerCamelCase : Optional[int] = f'''Warning: changing {name:{name_width}}''' for k, v in kwargs.items(): s += f''' {k}={v}''' setattr(lowercase__ , lowercase__ , lowercase__ ) logger.info(lowercase__ )
96
"""simple docstring""" import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler lowercase__ = 16 lowercase__ = 32 def _snake_case ( lowercase__ , lowercase__ = 16 , lowercase__ = "bert-base-cased" ): _lowerCamelCase : List[Any] = AutoTokenizer.from_pretrained(lowercase__ ) _lowerCamelCase : Tuple = load_dataset('glue' , 'mrpc' ) def tokenize_function(lowercase__ ): # max_length=None => use the model max length (it's actually the default) _lowerCamelCase : Union[str, Any] = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=lowercase__ , max_length=lowercase__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset _lowerCamelCase : int = datasets.map( lowercase__ , batched=lowercase__ , remove_columns=['idx', 'sentence1', 'sentence2'] , load_from_cache_file=lowercase__ ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _lowerCamelCase : Optional[int] = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(lowercase__ ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(lowercase__ , padding='max_length' , max_length=128 , return_tensors='pt' ) return tokenizer.pad(lowercase__ , padding='longest' , return_tensors='pt' ) # Instantiate dataloaders. _lowerCamelCase : List[str] = DataLoader( tokenized_datasets['train'] , shuffle=lowercase__ , collate_fn=lowercase__ , batch_size=lowercase__ ) _lowerCamelCase : int = DataLoader( tokenized_datasets['validation'] , shuffle=lowercase__ , collate_fn=lowercase__ , batch_size=lowercase__ ) return train_dataloader, eval_dataloader def _snake_case ( lowercase__ , lowercase__ ): # Initialize accelerator _lowerCamelCase : Optional[int] = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _lowerCamelCase : Optional[int] = config['lr'] _lowerCamelCase : Optional[int] = int(config['num_epochs'] ) _lowerCamelCase : Union[str, Any] = int(config['seed'] ) _lowerCamelCase : Optional[int] = int(config['batch_size'] ) _lowerCamelCase : Dict = args.model_name_or_path set_seed(lowercase__ ) _lowerCamelCase, _lowerCamelCase : Optional[int] = get_dataloaders(lowercase__ , lowercase__ , lowercase__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _lowerCamelCase : int = AutoModelForSequenceClassification.from_pretrained(lowercase__ , return_dict=lowercase__ ) # Instantiate optimizer _lowerCamelCase : Optional[int] = ( AdamW if accelerator.state.deepspeed_plugin is None or 'optimizer' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) _lowerCamelCase : Union[str, Any] = optimizer_cls(params=model.parameters() , lr=lowercase__ ) if accelerator.state.deepspeed_plugin is not None: _lowerCamelCase : str = accelerator.state.deepspeed_plugin.deepspeed_config[ 'gradient_accumulation_steps' ] else: _lowerCamelCase : Tuple = 1 _lowerCamelCase : List[Any] = (len(lowercase__ ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): _lowerCamelCase : Tuple = get_linear_schedule_with_warmup( optimizer=lowercase__ , num_warmup_steps=0 , num_training_steps=lowercase__ , ) else: _lowerCamelCase : Any = DummyScheduler(lowercase__ , total_num_steps=lowercase__ , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : Union[str, Any] = accelerator.prepare( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) # We need to keep track of how many total steps we have iterated over _lowerCamelCase : Union[str, Any] = 0 # We also need to keep track of the stating epoch so files are named properly _lowerCamelCase : Dict = 0 # Now we train the model _lowerCamelCase : Dict = evaluate.load('glue' , 'mrpc' ) _lowerCamelCase : Optional[int] = 0 _lowerCamelCase : str = {} for epoch in range(lowercase__ , lowercase__ ): model.train() for step, batch in enumerate(lowercase__ ): _lowerCamelCase : List[Any] = model(**lowercase__ ) _lowerCamelCase : int = outputs.loss _lowerCamelCase : Dict = loss / gradient_accumulation_steps accelerator.backward(lowercase__ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 model.eval() _lowerCamelCase : Union[str, Any] = 0 for step, batch in enumerate(lowercase__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _lowerCamelCase : Optional[int] = model(**lowercase__ ) _lowerCamelCase : Dict = outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times _lowerCamelCase, _lowerCamelCase : List[str] = accelerator.gather( (predictions, batch['labels']) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(lowercase__ ) - 1: _lowerCamelCase : Optional[Any] = predictions[: len(eval_dataloader.dataset ) - samples_seen] _lowerCamelCase : Dict = references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=lowercase__ , references=lowercase__ , ) _lowerCamelCase : List[Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f'''epoch {epoch}:''' , lowercase__ ) _lowerCamelCase : Tuple = eval_metric['accuracy'] if best_performance < eval_metric["accuracy"]: _lowerCamelCase : str = eval_metric['accuracy'] if args.performance_lower_bound is not None: assert ( args.performance_lower_bound <= best_performance ), f'''Best performance metric {best_performance} is lower than the lower bound {args.performance_lower_bound}''' accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , 'all_results.json' ) , 'w' ) as f: json.dump(lowercase__ , lowercase__ ) def _snake_case ( ): _lowerCamelCase : Any = argparse.ArgumentParser(description='Simple example of training script tracking peak GPU memory usage.' ) parser.add_argument( '--model_name_or_path' , type=lowercase__ , default='bert-base-cased' , help='Path to pretrained model or model identifier from huggingface.co/models.' , required=lowercase__ , ) parser.add_argument( '--output_dir' , type=lowercase__ , default='.' , help='Optional save directory where all checkpoint folders will be stored. Default is the current working directory.' , ) parser.add_argument( '--performance_lower_bound' , type=lowercase__ , default=lowercase__ , help='Optional lower bound for the performance metric. If set, the training will throw error when the performance metric drops below this value.' , ) parser.add_argument( '--num_epochs' , type=lowercase__ , default=3 , help='Number of train epochs.' , ) _lowerCamelCase : Optional[Any] = parser.parse_args() _lowerCamelCase : str = {'lr': 2E-5, 'num_epochs': args.num_epochs, 'seed': 42, 'batch_size': 16} training_function(lowercase__ , lowercase__ ) if __name__ == "__main__": main()
96
1
"""simple docstring""" import io import os import unicodedata from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = """▁""" lowercase__ = {"""vocab_file""": """vocab.txt""", """sentencepiece_model_ckpt""": """sentencepiece.bpe.model"""} lowercase__ = { """sentencepiece_model_file""": """sentencepiece.bpe.model""", """vocab_file""": """vocab.txt""", } lowercase__ = { """vocab_file""": { """ernie-m-base""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt""", """ernie-m-large""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt""", }, """sentencepiece_model_file""": { """ernie-m-base""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model""", """ernie-m-large""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model""", }, } lowercase__ = { """ernie-m-base""": 514, """ernie-m-large""": 514, } lowercase__ = { """ernie-m-base""": {"""do_lower_case""": False}, """ernie-m-large""": {"""do_lower_case""": False}, } class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = ["input_ids"] lowerCamelCase__ = VOCAB_FILES_NAMES lowerCamelCase__ = PRETRAINED_INIT_CONFIGURATION lowerCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ = RESOURCE_FILES_NAMES def __init__( self , lowercase , lowercase=None , lowercase=False , lowercase="utf8" , lowercase="[UNK]" , lowercase="[SEP]" , lowercase="[PAD]" , lowercase="[CLS]" , lowercase="[MASK]" , lowercase = None , **lowercase , ): # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. _lowerCamelCase : Optional[int] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=lowercase , unk_token=lowercase , sep_token=lowercase , pad_token=lowercase , cls_token=lowercase , mask_token=lowercase , vocab_file=lowercase , encoding=lowercase , sp_model_kwargs=self.sp_model_kwargs , **lowercase , ) _lowerCamelCase : str = do_lower_case _lowerCamelCase : Optional[Any] = sentencepiece_model_ckpt _lowerCamelCase : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowercase ) # to mimic paddlenlp.transformers.ernie_m.tokenizer.ErnieMTokenizer functioning if vocab_file is not None: _lowerCamelCase : Dict = self.load_vocab(filepath=lowercase ) else: _lowerCamelCase : Optional[int] = {self.sp_model.id_to_piece(lowercase ): id for id in range(self.sp_model.get_piece_size() )} _lowerCamelCase : int = {v: k for k, v in self.vocab.items()} def A_ ( self , lowercase ): if text is None: return None _lowerCamelCase : Tuple = self.tokenize(lowercase ) _lowerCamelCase, _lowerCamelCase : Any = '', [] for i, ch in enumerate(lowercase ): if ch in self.SP_CHAR_MAPPING: _lowerCamelCase : List[str] = self.SP_CHAR_MAPPING.get(lowercase ) else: _lowerCamelCase : Dict = unicodedata.normalize('NFKC' , lowercase ) if self.is_whitespace(lowercase ): continue normalized_text += ch char_mapping.extend([i] * len(lowercase ) ) _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : Union[str, Any] = normalized_text, [], 0 if self.do_lower_case: _lowerCamelCase : List[str] = text.lower() for token in split_tokens: if token[:1] == "▁": _lowerCamelCase : List[str] = token[1:] _lowerCamelCase : Union[str, Any] = text[offset:].index(lowercase ) + offset _lowerCamelCase : Optional[int] = start + len(lowercase ) token_mapping.append((char_mapping[start], char_mapping[end - 1] + 1) ) _lowerCamelCase : Optional[Any] = end return token_mapping @property def A_ ( self ): return len(self.vocab ) def A_ ( self ): return dict(self.vocab , **self.added_tokens_encoder ) def __getstate__( self ): _lowerCamelCase : str = self.__dict__.copy() _lowerCamelCase : Optional[int] = None return state def __setstate__( self , lowercase ): _lowerCamelCase : Union[str, Any] = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): _lowerCamelCase : List[str] = {} _lowerCamelCase : str = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.sentencepiece_model_ckpt ) def A_ ( self , lowercase ): return "".join((self.SP_CHAR_MAPPING.get(lowercase , lowercase ) for c in text) ) def A_ ( self , lowercase , lowercase=False , lowercase=64 , lowercase=0.1 ): if self.sp_model_kwargs.get('enable_sampling' ) is True: _lowerCamelCase : Optional[int] = True if self.sp_model_kwargs.get('alpha' ) is not None: _lowerCamelCase : Tuple = self.sp_model_kwargs.get('alpha' ) if self.sp_model_kwargs.get('nbest_size' ) is not None: _lowerCamelCase : Tuple = self.sp_model_kwargs.get('nbest_size' ) if not enable_sampling: _lowerCamelCase : Optional[Any] = self.sp_model.EncodeAsPieces(lowercase ) else: _lowerCamelCase : List[str] = self.sp_model.SampleEncodeAsPieces(lowercase , lowercase , lowercase ) _lowerCamelCase : Tuple = [] for pi, piece in enumerate(lowercase ): if piece == SPIECE_UNDERLINE: if not pieces[pi + 1].startswith(lowercase ) and pi != 0: new_pieces.append(lowercase ) continue else: continue _lowerCamelCase : Tuple = 0 for i, chunk in enumerate(lowercase ): if chunk == SPIECE_UNDERLINE: continue if self.is_ch_char(lowercase ) or self.is_punct(lowercase ): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) new_pieces.append(lowercase ) _lowerCamelCase : int = i + 1 elif chunk.isdigit() and i > 0 and not piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) _lowerCamelCase : Any = i elif not chunk.isdigit() and i > 0 and piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) _lowerCamelCase : Tuple = i if len(lowercase ) > lst_i: new_pieces.append(piece[lst_i:] ) return new_pieces def A_ ( self , lowercase ): _lowerCamelCase : Optional[int] = ''.join(lowercase ).replace(lowercase , ' ' ).strip() return out_string def A_ ( self , lowercase ): _lowerCamelCase : Union[str, Any] = self.convert_ids_to_tokens(lowercase ) _lowerCamelCase : Any = ''.join(lowercase ).replace(lowercase , ' ' ).strip() return out_string def A_ ( self , lowercase ): return self.vocab.get(lowercase , self.vocab.get(self.unk_token ) ) def A_ ( self , lowercase ): return self.reverse_vocab.get(lowercase , self.unk_token ) def A_ ( self , lowercase , lowercase=None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _lowerCamelCase : List[Any] = [self.cls_token_id] _lowerCamelCase : Optional[Any] = [self.sep_token_id] return _cls + token_ids_a + _sep + _sep + token_ids_a + _sep def A_ ( self , lowercase , lowercase=None ): if offset_mapping_a is None: return [(0, 0)] + offset_mapping_a + [(0, 0)] return [(0, 0)] + offset_mapping_a + [(0, 0), (0, 0)] + offset_mapping_a + [(0, 0)] def A_ ( self , lowercase , lowercase=None , lowercase=False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( 'You should not supply a second sequence if the provided sequence of ' 'ids is already formatted with special tokens for the model.' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(lowercase )) + [1, 1] + ([0] * len(lowercase )) + [1] return [1] + ([0] * len(lowercase )) + [1] def A_ ( self , lowercase , lowercase = None ): # called when `add_special_tokens` is True, so align with `build_inputs_with_special_tokens` method if token_ids_a is None: # [CLS] X [SEP] return (len(lowercase ) + 2) * [0] # [CLS] A [SEP] [SEP] B [SEP] return [0] * (len(lowercase ) + 1) + [1] * (len(lowercase ) + 3) def A_ ( self , lowercase ): if "\u4e00" <= char <= "\u9fff": return True return False def A_ ( self , lowercase ): if ("a" <= char <= "z") or ("A" <= char <= "Z"): return True return False def A_ ( self , lowercase ): if char in ",;:.?!~,;:。?!《》【】": return True return False def A_ ( self , lowercase ): if char == " " or char == "\t" or char == "\n" or char == "\r": return True if len(lowercase ) == 1: _lowerCamelCase : Tuple = unicodedata.category(lowercase ) if cat == "Zs": return True return False def A_ ( self , lowercase ): _lowerCamelCase : Tuple = {} with io.open(lowercase , 'r' , encoding='utf-8' ) as f: for index, line in enumerate(lowercase ): _lowerCamelCase : int = line.rstrip('\n' ) _lowerCamelCase : Optional[Any] = int(lowercase ) return token_to_idx def A_ ( self , lowercase , lowercase = None ): _lowerCamelCase : Union[str, Any] = 0 if os.path.isdir(lowercase ): _lowerCamelCase : List[Any] = os.path.join( lowercase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) else: _lowerCamelCase : Optional[int] = (filename_prefix + '-' if filename_prefix else '') + save_directory with open(lowercase , 'w' , encoding='utf-8' ) as writer: for token, token_index in sorted(self.vocab.items() , key=lambda lowercase : kv[1] ): if index != token_index: logger.warning( F'''Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.''' ' Please check that the vocabulary is not corrupted!' ) _lowerCamelCase : Optional[Any] = token_index writer.write(token + '\n' ) index += 1 _lowerCamelCase : List[str] = os.path.join(lowercase , 'sentencepiece.bpe.model' ) with open(lowercase , 'wb' ) as fi: _lowerCamelCase : Dict = self.sp_model.serialized_model_proto() fi.write(lowercase ) return (vocab_file,)
96
"""simple docstring""" from __future__ import annotations import copy import tempfile import unittest from transformers import CONFIG_MAPPING, AutoConfig, BertConfig, GPTaConfig, TaConfig, TapasConfig, is_tf_available from transformers.testing_utils import ( DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tensorflow_probability, require_tf, slow, ) from ..bert.test_modeling_bert import BertModelTester if is_tf_available(): from transformers import ( TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelForTableQuestionAnswering, TFAutoModelForTokenClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFFunnelBaseModel, TFFunnelModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, TFTapasForQuestionAnswering, ) from transformers.models.auto.modeling_tf_auto import ( TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_MAPPING, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.tapas.modeling_tf_tapas import TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """new-model""" if is_tf_available(): class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = NewModelConfig @require_tf class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' @slow def A_ ( self ): _lowerCamelCase : List[str] = 'bert-base-cased' _lowerCamelCase : Union[str, Any] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Union[str, Any] = TFAutoModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): _lowerCamelCase : List[str] = 'bert-base-cased' _lowerCamelCase : int = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : int = TFAutoModelForPreTraining.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Dict = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : int = TFAutoModelForCausalLM.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : str = TFAutoModelForCausalLM.from_pretrained(lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : List[Any] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : str = TFAutoModelWithLMHead.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Tuple = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Optional[int] = TFAutoModelForMaskedLM.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : Tuple = TFAutoModelForMaskedLM.from_pretrained(lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Optional[int] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Tuple = TFAutoModelForSeqaSeqLM.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : Tuple = TFAutoModelForSeqaSeqLM.from_pretrained(lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: _lowerCamelCase : str = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Union[str, Any] = TFAutoModelForSequenceClassification.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow def A_ ( self ): # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: _lowerCamelCase : Optional[Any] = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : List[str] = TFAutoModelForQuestionAnswering.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) @slow @require_tensorflow_probability def A_ ( self ): for model_name in TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST[5:6]: _lowerCamelCase : Dict = AutoConfig.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Dict = TFAutoModelForTableQuestionAnswering.from_pretrained(lowercase ) _lowerCamelCase, _lowerCamelCase : List[Any] = TFAutoModelForTableQuestionAnswering.from_pretrained( lowercase , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertIsInstance(lowercase , lowercase ) def A_ ( self ): _lowerCamelCase : int = TFAutoModelWithLMHead.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowercase ) , 14410 ) def A_ ( self ): _lowerCamelCase : Any = TFAutoModelWithLMHead.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowercase ) , 14410 ) def A_ ( self ): # For the auto model mapping, FunnelConfig has two models: FunnelModel and FunnelBaseModel _lowerCamelCase : List[str] = TFAutoModel.from_pretrained('sgugger/funnel-random-tiny' ) self.assertIsInstance(lowercase , lowercase ) _lowerCamelCase : Optional[int] = copy.deepcopy(model.config ) _lowerCamelCase : Dict = ['FunnelBaseModel'] _lowerCamelCase : List[Any] = TFAutoModel.from_config(lowercase ) self.assertIsInstance(lowercase , lowercase ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowercase ) _lowerCamelCase : Tuple = TFAutoModel.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) def A_ ( self ): try: AutoConfig.register('new-model' , lowercase ) _lowerCamelCase : Tuple = [ TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSequenceClassification, TFAutoModelForTokenClassification, ] for auto_class in auto_classes: with self.subTest(auto_class.__name__ ): # Wrong config class will raise an error with self.assertRaises(lowercase ): auto_class.register(lowercase , lowercase ) auto_class.register(lowercase , lowercase ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(lowercase ): auto_class.register(lowercase , lowercase ) # Now that the config is registered, it can be used as any other config with the auto-API _lowerCamelCase : Optional[Any] = BertModelTester(self ).get_config() _lowerCamelCase : Dict = NewModelConfig(**tiny_config.to_dict() ) _lowerCamelCase : int = auto_class.from_config(lowercase ) self.assertIsInstance(lowercase , lowercase ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(lowercase ) _lowerCamelCase : List[Any] = auto_class.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"] for mapping in ( TF_MODEL_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, ): if NewModelConfig in mapping._extra_content: del mapping._extra_content[NewModelConfig] def A_ ( self ): with self.assertRaisesRegex( lowercase , 'bert-base is not a local folder and is not a valid model identifier' ): _lowerCamelCase : Union[str, Any] = TFAutoModel.from_pretrained('bert-base' ) def A_ ( self ): with self.assertRaisesRegex( lowercase , r'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): _lowerCamelCase : str = TFAutoModel.from_pretrained(lowercase , revision='aaaaaa' ) def A_ ( self ): with self.assertRaisesRegex( lowercase , 'hf-internal-testing/config-no-model does not appear to have a file named pytorch_model.bin' , ): _lowerCamelCase : Tuple = TFAutoModel.from_pretrained('hf-internal-testing/config-no-model' ) def A_ ( self ): with self.assertRaisesRegex(lowercase , 'Use `from_pt=True` to load this model' ): _lowerCamelCase : Tuple = TFAutoModel.from_pretrained('hf-internal-testing/tiny-bert-pt-only' ) def A_ ( self ): # Make sure we have cached the model. _lowerCamelCase : Optional[int] = TFAutoModel.from_pretrained('hf-internal-testing/tiny-random-bert' ) with RequestCounter() as counter: _lowerCamelCase : Optional[int] = TFAutoModel.from_pretrained('hf-internal-testing/tiny-random-bert' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 ) # With a sharded checkpoint _lowerCamelCase : int = TFAutoModel.from_pretrained('ArthurZ/tiny-random-bert-sharded' ) with RequestCounter() as counter: _lowerCamelCase : List[Any] = TFAutoModel.from_pretrained('ArthurZ/tiny-random-bert-sharded' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 )
96
1
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyInpaintPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.text_encoder import MCLIPConfig, MultilingualCLIP from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class lowerCAmelCase__ ( lowercase, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = KandinskyInpaintPipeline lowerCamelCase__ = ["""prompt""", """image_embeds""", """negative_image_embeds""", """image""", """mask_image"""] lowerCamelCase__ = [ """prompt""", """negative_prompt""", """image_embeds""", """negative_image_embeds""", """image""", """mask_image""", ] lowerCamelCase__ = [ """generator""", """height""", """width""", """latents""", """guidance_scale""", """negative_prompt""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] lowerCamelCase__ = False @property def A_ ( self ): return 32 @property def A_ ( self ): return 32 @property def A_ ( self ): return self.time_input_dim @property def A_ ( self ): return self.time_input_dim * 4 @property def A_ ( self ): return 100 @property def A_ ( self ): _lowerCamelCase : Tuple = XLMRobertaTokenizerFast.from_pretrained('YiYiXu/tiny-random-mclip-base' ) return tokenizer @property def A_ ( self ): torch.manual_seed(0 ) _lowerCamelCase : str = MCLIPConfig( numDims=self.cross_attention_dim , transformerDimensions=self.text_embedder_hidden_size , hidden_size=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=1005 , ) _lowerCamelCase : str = MultilingualCLIP(lowercase ) _lowerCamelCase : Optional[Any] = text_encoder.eval() return text_encoder @property def A_ ( self ): torch.manual_seed(0 ) _lowerCamelCase : List[str] = { 'in_channels': 9, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'text_image', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'text_image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } _lowerCamelCase : Union[str, Any] = UNetaDConditionModel(**lowercase ) return model @property def A_ ( self ): return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def A_ ( self ): torch.manual_seed(0 ) _lowerCamelCase : str = VQModel(**self.dummy_movq_kwargs ) return model def A_ ( self ): _lowerCamelCase : Optional[int] = self.dummy_text_encoder _lowerCamelCase : Optional[Any] = self.dummy_tokenizer _lowerCamelCase : Optional[int] = self.dummy_unet _lowerCamelCase : Tuple = self.dummy_movq _lowerCamelCase : Dict = DDIMScheduler( num_train_timesteps=1000 , beta_schedule='linear' , beta_start=0.0_00_85 , beta_end=0.0_12 , clip_sample=lowercase , set_alpha_to_one=lowercase , steps_offset=1 , prediction_type='epsilon' , thresholding=lowercase , ) _lowerCamelCase : List[str] = { 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def A_ ( self , lowercase , lowercase=0 ): _lowerCamelCase : List[Any] = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(lowercase ) ).to(lowercase ) _lowerCamelCase : List[str] = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(seed + 1 ) ).to(lowercase ) # create init_image _lowerCamelCase : int = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase ) ).to(lowercase ) _lowerCamelCase : Union[str, Any] = image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowerCamelCase : List[str] = Image.fromarray(np.uinta(lowercase ) ).convert('RGB' ).resize((256, 256) ) # create mask _lowerCamelCase : Tuple = np.ones((64, 64) , dtype=np.floataa ) _lowerCamelCase : Optional[int] = 0 if str(lowercase ).startswith('mps' ): _lowerCamelCase : List[Any] = torch.manual_seed(lowercase ) else: _lowerCamelCase : List[Any] = torch.Generator(device=lowercase ).manual_seed(lowercase ) _lowerCamelCase : str = { 'prompt': 'horse', 'image': init_image, 'mask_image': mask, 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'generator': generator, 'height': 64, 'width': 64, 'num_inference_steps': 2, 'guidance_scale': 4.0, 'output_type': 'np', } return inputs def A_ ( self ): _lowerCamelCase : List[Any] = 'cpu' _lowerCamelCase : Tuple = self.get_dummy_components() _lowerCamelCase : List[str] = self.pipeline_class(**lowercase ) _lowerCamelCase : Optional[int] = pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : int = pipe(**self.get_dummy_inputs(lowercase ) ) _lowerCamelCase : List[Any] = output.images _lowerCamelCase : Dict = pipe( **self.get_dummy_inputs(lowercase ) , return_dict=lowercase , )[0] _lowerCamelCase : Optional[int] = image[0, -3:, -3:, -1] _lowerCamelCase : str = image_from_tuple[0, -3:, -3:, -1] print(F'''image.shape {image.shape}''' ) assert image.shape == (1, 64, 64, 3) _lowerCamelCase : List[Any] = np.array( [0.8_32_69_19, 0.73_79_04_67, 0.20_91_85_81, 0.9_30_96_12, 0.5_51_17_91, 0.43_71_33_28, 0.5_51_33_21, 0.49_92_29_34, 0.59_49_77_86] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), F''' expected_slice {expected_slice}, but got {image_slice.flatten()}''' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), F''' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}''' def A_ ( self ): super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def A_ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A_ ( self ): _lowerCamelCase : List[str] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/kandinsky_inpaint_cat_with_hat_fp16.npy' ) _lowerCamelCase : Optional[int] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/cat.png' ) _lowerCamelCase : str = np.ones((768, 768) , dtype=np.floataa ) _lowerCamelCase : Any = 0 _lowerCamelCase : Any = 'a hat' _lowerCamelCase : Optional[int] = KandinskyPriorPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-1-prior' , torch_dtype=torch.floataa ) pipe_prior.to(lowercase ) _lowerCamelCase : List[Any] = KandinskyInpaintPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-1-inpaint' , torch_dtype=torch.floataa ) _lowerCamelCase : Optional[int] = pipeline.to(lowercase ) pipeline.set_progress_bar_config(disable=lowercase ) _lowerCamelCase : Optional[Any] = torch.Generator(device='cpu' ).manual_seed(0 ) _lowerCamelCase, _lowerCamelCase : Optional[Any] = pipe_prior( lowercase , generator=lowercase , num_inference_steps=5 , negative_prompt='' , ).to_tuple() _lowerCamelCase : int = pipeline( lowercase , image=lowercase , mask_image=lowercase , image_embeds=lowercase , negative_image_embeds=lowercase , generator=lowercase , num_inference_steps=100 , height=768 , width=768 , output_type='np' , ) _lowerCamelCase : Dict = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(lowercase , lowercase )
96
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ = {"""configuration_ibert""": ["""IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """IBertConfig""", """IBertOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ """IBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """IBertForMaskedLM""", """IBertForMultipleChoice""", """IBertForQuestionAnswering""", """IBertForSequenceClassification""", """IBertForTokenClassification""", """IBertModel""", """IBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
96
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowercase__ = {"""configuration_unispeech""": ["""UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP""", """UniSpeechConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ """UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST""", """UniSpeechForCTC""", """UniSpeechForPreTraining""", """UniSpeechForSequenceClassification""", """UniSpeechModel""", """UniSpeechPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_unispeech import UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP, UniSpeechConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_unispeech import ( UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST, UniSpeechForCTC, UniSpeechForPreTraining, UniSpeechForSequenceClassification, UniSpeechModel, UniSpeechPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
96
"""simple docstring""" import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : Tuple = f'''{sampling_rate}''' _lowerCamelCase : str = '1' _lowerCamelCase : str = 'f32le' _lowerCamelCase : Union[str, Any] = [ 'ffmpeg', '-i', 'pipe:0', '-ac', ac, '-ar', ar, '-f', format_for_conversion, '-hide_banner', '-loglevel', 'quiet', 'pipe:1', ] try: with subprocess.Popen(lowercase__ , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: _lowerCamelCase : str = ffmpeg_process.communicate(lowercase__ ) except FileNotFoundError as error: raise ValueError('ffmpeg was not found but is required to load audio files from filename' ) from error _lowerCamelCase : List[Any] = output_stream[0] _lowerCamelCase : Tuple = np.frombuffer(lowercase__ , np.floataa ) if audio.shape[0] == 0: raise ValueError('Malformed soundfile' ) return audio def _snake_case ( lowercase__ , lowercase__ , lowercase__ = "f32le" , ): _lowerCamelCase : Optional[Any] = f'''{sampling_rate}''' _lowerCamelCase : List[str] = '1' if format_for_conversion == "s16le": _lowerCamelCase : List[str] = 2 elif format_for_conversion == "f32le": _lowerCamelCase : List[Any] = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) _lowerCamelCase : Dict = platform.system() if system == "Linux": _lowerCamelCase : Optional[int] = 'alsa' _lowerCamelCase : Optional[Any] = 'default' elif system == "Darwin": _lowerCamelCase : Optional[int] = 'avfoundation' _lowerCamelCase : Any = ':0' elif system == "Windows": _lowerCamelCase : Tuple = 'dshow' _lowerCamelCase : Tuple = 'default' _lowerCamelCase : Optional[int] = [ 'ffmpeg', '-f', format_, '-i', input_, '-ac', ac, '-ar', ar, '-f', format_for_conversion, '-fflags', 'nobuffer', '-hide_banner', '-loglevel', 'quiet', 'pipe:1', ] _lowerCamelCase : Tuple = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample _lowerCamelCase : List[Any] = _ffmpeg_stream(lowercase__ , lowercase__ ) for item in iterator: yield item def _snake_case ( lowercase__ , lowercase__ , lowercase__ = None , lowercase__ = None , lowercase__ = "f32le" , ): if stream_chunk_s is not None: _lowerCamelCase : int = stream_chunk_s else: _lowerCamelCase : Optional[Any] = chunk_length_s _lowerCamelCase : Optional[Any] = ffmpeg_microphone(lowercase__ , lowercase__ , format_for_conversion=lowercase__ ) if format_for_conversion == "s16le": _lowerCamelCase : List[str] = np.intaa _lowerCamelCase : str = 2 elif format_for_conversion == "f32le": _lowerCamelCase : Any = np.floataa _lowerCamelCase : List[Any] = 4 else: raise ValueError(f'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) if stride_length_s is None: _lowerCamelCase : Union[str, Any] = chunk_length_s / 6 _lowerCamelCase : Optional[int] = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(lowercase__ , (int, float) ): _lowerCamelCase : Any = [stride_length_s, stride_length_s] _lowerCamelCase : Tuple = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample _lowerCamelCase : Optional[Any] = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample _lowerCamelCase : List[Any] = datetime.datetime.now() _lowerCamelCase : Optional[int] = datetime.timedelta(seconds=lowercase__ ) for item in chunk_bytes_iter(lowercase__ , lowercase__ , stride=(stride_left, stride_right) , stream=lowercase__ ): # Put everything back in numpy scale _lowerCamelCase : List[Any] = np.frombuffer(item['raw'] , dtype=lowercase__ ) _lowerCamelCase : int = ( item['stride'][0] // size_of_sample, item['stride'][1] // size_of_sample, ) _lowerCamelCase : Optional[int] = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def _snake_case ( lowercase__ , lowercase__ , lowercase__ , lowercase__ = False ): _lowerCamelCase : int = B'' _lowerCamelCase, _lowerCamelCase : Dict = stride if stride_left + stride_right >= chunk_len: raise ValueError( f'''Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}''' ) _lowerCamelCase : str = 0 for raw in iterator: acc += raw if stream and len(lowercase__ ) < chunk_len: _lowerCamelCase : Optional[int] = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(lowercase__ ) >= chunk_len: # We are flushing the accumulator _lowerCamelCase : str = (_stride_left, stride_right) _lowerCamelCase : str = {'raw': acc[:chunk_len], 'stride': stride} if stream: _lowerCamelCase : List[Any] = False yield item _lowerCamelCase : Optional[Any] = stride_left _lowerCamelCase : str = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(lowercase__ ) > stride_left: _lowerCamelCase : Optional[Any] = {'raw': acc, 'stride': (_stride_left, 0)} if stream: _lowerCamelCase : Tuple = False yield item def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : int = 2**24 # 16Mo try: with subprocess.Popen(lowercase__ , stdout=subprocess.PIPE , bufsize=lowercase__ ) as ffmpeg_process: while True: _lowerCamelCase : Optional[Any] = ffmpeg_process.stdout.read(lowercase__ ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError('ffmpeg was not found but is required to stream audio files from filename' ) from error
96
1
"""simple docstring""" import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { """Salesforce/blip-vqa-base""": """https://huggingface.co/Salesforce/blip-vqa-base/resolve/main/config.json""", """Salesforce/blip-vqa-capfit-large""": ( """https://huggingface.co/Salesforce/blip-vqa-base-capfit/resolve/main/config.json""" ), """Salesforce/blip-image-captioning-base""": ( """https://huggingface.co/Salesforce/blip-image-captioning-base/resolve/main/config.json""" ), """Salesforce/blip-image-captioning-large""": ( """https://huggingface.co/Salesforce/blip-image-captioning-large/resolve/main/config.json""" ), """Salesforce/blip-itm-base-coco""": """https://huggingface.co/Salesforce/blip-itm-base-coco/resolve/main/config.json""", """Salesforce/blip-itm-large-coco""": """https://huggingface.co/Salesforce/blip-itm-large-coco/resolve/main/config.json""", """Salesforce/blip-itm-base-flikr""": """https://huggingface.co/Salesforce/blip-itm-base-flikr/resolve/main/config.json""", """Salesforce/blip-itm-large-flikr""": ( """https://huggingface.co/Salesforce/blip-itm-large-flikr/resolve/main/config.json""" ), } class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """blip_text_model""" def __init__( self , lowercase=30524 , lowercase=768 , lowercase=768 , lowercase=3072 , lowercase=768 , lowercase=12 , lowercase=8 , lowercase=512 , lowercase="gelu" , lowercase=1E-12 , lowercase=0.0 , lowercase=0.0 , lowercase=0.02 , lowercase=30522 , lowercase=2 , lowercase=0 , lowercase=102 , lowercase=True , lowercase=True , **lowercase , ): super().__init__( pad_token_id=lowercase , bos_token_id=lowercase , eos_token_id=lowercase , sep_token_id=lowercase , **lowercase , ) _lowerCamelCase : List[str] = vocab_size _lowerCamelCase : Tuple = hidden_size _lowerCamelCase : Optional[int] = encoder_hidden_size _lowerCamelCase : List[Any] = intermediate_size _lowerCamelCase : Optional[Any] = projection_dim _lowerCamelCase : Tuple = hidden_dropout_prob _lowerCamelCase : Tuple = num_hidden_layers _lowerCamelCase : int = num_attention_heads _lowerCamelCase : Optional[Any] = max_position_embeddings _lowerCamelCase : Optional[int] = layer_norm_eps _lowerCamelCase : List[Any] = hidden_act _lowerCamelCase : Tuple = initializer_range _lowerCamelCase : int = attention_probs_dropout_prob _lowerCamelCase : List[str] = is_decoder _lowerCamelCase : Optional[Any] = use_cache @classmethod def A_ ( cls , lowercase , **lowercase ): cls._set_token_in_kwargs(lowercase ) _lowerCamelCase, _lowerCamelCase : int = cls.get_config_dict(lowercase , **lowercase ) # get the text config dict if we are loading from BlipConfig if config_dict.get('model_type' ) == "blip": _lowerCamelCase : Union[str, Any] = config_dict['text_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(lowercase , **lowercase ) class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """blip_vision_model""" def __init__( self , lowercase=768 , lowercase=3072 , lowercase=512 , lowercase=12 , lowercase=12 , lowercase=384 , lowercase=16 , lowercase="gelu" , lowercase=1E-5 , lowercase=0.0 , lowercase=1E-10 , **lowercase , ): super().__init__(**lowercase ) _lowerCamelCase : Union[str, Any] = hidden_size _lowerCamelCase : List[str] = intermediate_size _lowerCamelCase : Any = projection_dim _lowerCamelCase : List[Any] = num_hidden_layers _lowerCamelCase : Any = num_attention_heads _lowerCamelCase : List[Any] = patch_size _lowerCamelCase : Union[str, Any] = image_size _lowerCamelCase : Optional[int] = initializer_range _lowerCamelCase : Optional[int] = attention_dropout _lowerCamelCase : List[Any] = layer_norm_eps _lowerCamelCase : Tuple = hidden_act @classmethod def A_ ( cls , lowercase , **lowercase ): cls._set_token_in_kwargs(lowercase ) _lowerCamelCase, _lowerCamelCase : List[Any] = cls.get_config_dict(lowercase , **lowercase ) # get the vision config dict if we are loading from BlipConfig if config_dict.get('model_type' ) == "blip": _lowerCamelCase : List[Any] = config_dict['vision_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(lowercase , **lowercase ) class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """blip""" lowerCamelCase__ = True def __init__( self , lowercase=None , lowercase=None , lowercase=512 , lowercase=2.65_92 , lowercase=256 , **lowercase , ): super().__init__(**lowercase ) if text_config is None: _lowerCamelCase : int = {} logger.info('`text_config` is `None`. Initializing the `BlipTextConfig` with default values.' ) if vision_config is None: _lowerCamelCase : Tuple = {} logger.info('`vision_config` is `None`. Initializing the `BlipVisionConfig` with default values.' ) _lowerCamelCase : Dict = BlipTextConfig(**lowercase ) _lowerCamelCase : Optional[Any] = BlipVisionConfig(**lowercase ) _lowerCamelCase : int = self.vision_config.hidden_size _lowerCamelCase : List[str] = projection_dim _lowerCamelCase : List[Any] = logit_scale_init_value _lowerCamelCase : str = 1.0 _lowerCamelCase : List[Any] = 0.02 _lowerCamelCase : List[Any] = image_text_hidden_size @classmethod def A_ ( cls , lowercase , lowercase , **lowercase ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **lowercase ) def A_ ( self ): _lowerCamelCase : Dict = copy.deepcopy(self.__dict__ ) _lowerCamelCase : Optional[Any] = self.text_config.to_dict() _lowerCamelCase : str = self.vision_config.to_dict() _lowerCamelCase : Optional[Any] = self.__class__.model_type return output
96
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = {"""ctrl""": """https://huggingface.co/ctrl/resolve/main/config.json"""} class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """ctrl""" lowerCamelCase__ = ["""past_key_values"""] lowerCamelCase__ = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowercase=246534 , lowercase=256 , lowercase=1280 , lowercase=8192 , lowercase=48 , lowercase=16 , lowercase=0.1 , lowercase=0.1 , lowercase=1E-6 , lowercase=0.02 , lowercase=True , **lowercase , ): _lowerCamelCase : Any = vocab_size _lowerCamelCase : Dict = n_positions _lowerCamelCase : Optional[int] = n_embd _lowerCamelCase : str = n_layer _lowerCamelCase : Union[str, Any] = n_head _lowerCamelCase : Any = dff _lowerCamelCase : int = resid_pdrop _lowerCamelCase : Dict = embd_pdrop _lowerCamelCase : Union[str, Any] = layer_norm_epsilon _lowerCamelCase : Tuple = initializer_range _lowerCamelCase : str = use_cache super().__init__(**lowercase )
96
1
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require_tokenizers class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' @slow def A_ ( self ): _lowerCamelCase : int = AutoModelForSeqaSeqLM.from_pretrained('google/mt5-small' , return_dict=lowercase ).to(lowercase ) _lowerCamelCase : Union[str, Any] = AutoTokenizer.from_pretrained('google/mt5-small' ) _lowerCamelCase : Optional[Any] = tokenizer('Hello there' , return_tensors='pt' ).input_ids _lowerCamelCase : Dict = tokenizer('Hi I am' , return_tensors='pt' ).input_ids _lowerCamelCase : Dict = model(input_ids.to(lowercase ) , labels=labels.to(lowercase ) ).loss _lowerCamelCase : List[Any] = -(labels.shape[-1] * loss.item()) _lowerCamelCase : Union[str, Any] = -84.91_27 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
96
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator from typing import Any class lowerCAmelCase__ : '''simple docstring''' def __init__( self , lowercase ): _lowerCamelCase : Any = data _lowerCamelCase : Node | None = None class lowerCAmelCase__ : '''simple docstring''' def __init__( self ): _lowerCamelCase : str = None _lowerCamelCase : str = None def __iter__( self ): _lowerCamelCase : List[str] = self.head while self.head: yield node.data _lowerCamelCase : Optional[int] = node.next if node == self.head: break def __len__( self ): return sum(1 for _ in self ) def __repr__( self ): return "->".join(str(lowercase ) for item in iter(self ) ) def A_ ( self , lowercase ): self.insert_nth(len(self ) , lowercase ) def A_ ( self , lowercase ): self.insert_nth(0 , lowercase ) def A_ ( self , lowercase , lowercase ): if index < 0 or index > len(self ): raise IndexError('list index out of range.' ) _lowerCamelCase : List[Any] = Node(lowercase ) if self.head is None: _lowerCamelCase : str = new_node # first node points itself _lowerCamelCase : Union[str, Any] = new_node elif index == 0: # insert at head _lowerCamelCase : List[str] = self.head _lowerCamelCase : str = new_node else: _lowerCamelCase : Union[str, Any] = self.head for _ in range(index - 1 ): _lowerCamelCase : List[Any] = temp.next _lowerCamelCase : Union[str, Any] = temp.next _lowerCamelCase : List[str] = new_node if index == len(self ) - 1: # insert at tail _lowerCamelCase : Any = new_node def A_ ( self ): return self.delete_nth(0 ) def A_ ( self ): return self.delete_nth(len(self ) - 1 ) def A_ ( self , lowercase = 0 ): if not 0 <= index < len(self ): raise IndexError('list index out of range.' ) _lowerCamelCase : Any = self.head if self.head == self.tail: # just one node _lowerCamelCase : List[str] = None elif index == 0: # delete head node _lowerCamelCase : List[str] = self.tail.next.next _lowerCamelCase : Optional[int] = self.head.next else: _lowerCamelCase : Dict = self.head for _ in range(index - 1 ): _lowerCamelCase : List[Any] = temp.next _lowerCamelCase : int = temp.next _lowerCamelCase : Optional[int] = temp.next.next if index == len(self ) - 1: # delete at tail _lowerCamelCase : List[Any] = temp return delete_node.data def A_ ( self ): return len(self ) == 0 def _snake_case ( ): _lowerCamelCase : Union[str, Any] = CircularLinkedList() assert len(lowercase__ ) == 0 assert circular_linked_list.is_empty() is True assert str(lowercase__ ) == "" try: circular_linked_list.delete_front() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_tail() raise AssertionError # This should not happen except IndexError: assert True # This should happen try: circular_linked_list.delete_nth(-1 ) raise AssertionError except IndexError: assert True try: circular_linked_list.delete_nth(0 ) raise AssertionError except IndexError: assert True assert circular_linked_list.is_empty() is True for i in range(5 ): assert len(lowercase__ ) == i circular_linked_list.insert_nth(lowercase__ , i + 1 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 6 ) ) circular_linked_list.insert_tail(6 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 7 ) ) circular_linked_list.insert_head(0 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(0 , 7 ) ) assert circular_linked_list.delete_front() == 0 assert circular_linked_list.delete_tail() == 6 assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 6 ) ) assert circular_linked_list.delete_nth(2 ) == 3 circular_linked_list.insert_nth(2 , 3 ) assert str(lowercase__ ) == "->".join(str(lowercase__ ) for i in range(1 , 6 ) ) assert circular_linked_list.is_empty() is False if __name__ == "__main__": import doctest doctest.testmod()
96
1