code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __lowercase : Optional[Any] = logging.get_logger(__name__) __lowercase : Tuple = { 'microsoft/markuplm-base': 'https://huggingface.co/microsoft/markuplm-base/resolve/main/config.json', 'microsoft/markuplm-large': 'https://huggingface.co/microsoft/markuplm-large/resolve/main/config.json', } class __UpperCamelCase ( lowerCAmelCase_ ): A_ = "markuplm" def __init__( self , __a=3_0522 , __a=768 , __a=12 , __a=12 , __a=3072 , __a="gelu" , __a=0.1 , __a=0.1 , __a=512 , __a=2 , __a=0.02 , __a=1E-1_2 , __a=0 , __a=0 , __a=2 , __a=256 , __a=1024 , __a=216 , __a=1001 , __a=32 , __a=50 , __a="absolute" , __a=True , __a=None , **__a , ): '''simple docstring''' super().__init__( pad_token_id=__a , bos_token_id=__a , eos_token_id=__a , **__a , ) __a : List[str] = vocab_size __a : Optional[Any] = hidden_size __a : Optional[int] = num_hidden_layers __a : Optional[Any] = num_attention_heads __a : Optional[int] = hidden_act __a : Union[str, Any] = intermediate_size __a : str = hidden_dropout_prob __a : List[Any] = attention_probs_dropout_prob __a : Dict = max_position_embeddings __a : Optional[Any] = type_vocab_size __a : Union[str, Any] = initializer_range __a : Optional[Any] = layer_norm_eps __a : Optional[Any] = position_embedding_type __a : Optional[int] = use_cache __a : Optional[Any] = classifier_dropout # additional properties __a : int = max_depth __a : Union[str, Any] = max_xpath_tag_unit_embeddings __a : str = max_xpath_subs_unit_embeddings __a : Optional[Any] = tag_pad_id __a : Dict = subs_pad_id __a : List[str] = xpath_unit_hidden_size
27
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ : Dict = logging.get_logger(__name__) UpperCAmelCase_ : Union[str, Any] = { 'google/switch-base-8': 'https://huggingface.co/google/switch-base-8/blob/main/config.json', } class SCREAMING_SNAKE_CASE__ ( lowercase__ ): snake_case__ : List[str] = '''switch_transformers''' snake_case__ : Optional[int] = ['''past_key_values'''] snake_case__ : Optional[Any] = {'''hidden_size''': '''d_model''', '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers'''} def __init__( self : Dict , SCREAMING_SNAKE_CASE__ : Optional[int]=3_2_1_2_8 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=7_6_8 , SCREAMING_SNAKE_CASE__ : Optional[Any]=6_4 , SCREAMING_SNAKE_CASE__ : List[str]=2_0_4_8 , SCREAMING_SNAKE_CASE__ : Dict=6_4 , SCREAMING_SNAKE_CASE__ : List[Any]=1_2 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=3 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=1_2 , SCREAMING_SNAKE_CASE__ : str=3 , SCREAMING_SNAKE_CASE__ : Tuple=1_2 , SCREAMING_SNAKE_CASE__ : Tuple=8 , SCREAMING_SNAKE_CASE__ : Dict=False , SCREAMING_SNAKE_CASE__ : Union[str, Any]=0.01 , SCREAMING_SNAKE_CASE__ : str="float32" , SCREAMING_SNAKE_CASE__ : str=False , SCREAMING_SNAKE_CASE__ : Union[str, Any]=3_2 , SCREAMING_SNAKE_CASE__ : Dict=1_2_8 , SCREAMING_SNAKE_CASE__ : Dict=0.1 , SCREAMING_SNAKE_CASE__ : Dict=1E-6 , SCREAMING_SNAKE_CASE__ : Dict=0.001 , SCREAMING_SNAKE_CASE__ : Any=0.001 , SCREAMING_SNAKE_CASE__ : Optional[int]=1.0 , SCREAMING_SNAKE_CASE__ : Any="relu" , SCREAMING_SNAKE_CASE__ : List[Any]=True , SCREAMING_SNAKE_CASE__ : Optional[Any]=False , SCREAMING_SNAKE_CASE__ : Dict=True , SCREAMING_SNAKE_CASE__ : Union[str, Any]=0 , SCREAMING_SNAKE_CASE__ : Optional[int]=1 , **SCREAMING_SNAKE_CASE__ : Dict , ) -> Optional[Any]: a_ : Optional[int] = vocab_size a_ : List[str] = d_model a_ : Tuple = d_kv a_ : Optional[Any] = d_ff a_ : List[Any] = num_sparse_encoder_layers a_ : Any = num_layers a_ : str = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry a_ : List[Any] = num_sparse_decoder_layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_encoder_layers > 0: a_ : Optional[int] = self.num_layers // self.num_sparse_encoder_layers else: a_ : List[Any] = self.num_layers # HACK: this will create 0 sparse layers # This tells us, each how many encoder layer we'll have to set a sparse layer. if self.num_sparse_decoder_layers > 0: a_ : Union[str, Any] = self.num_decoder_layers // self.num_sparse_decoder_layers else: a_ : List[str] = self.num_decoder_layers # HACK: this will create 0 sparse layers a_ : Dict = num_heads a_ : str = num_experts a_ : Any = expert_capacity a_ : List[Any] = router_bias a_ : str = router_jitter_noise if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(F"""`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}""" ) a_ : Optional[int] = router_dtype a_ : int = router_ignore_padding_tokens a_ : Any = relative_attention_num_buckets a_ : List[str] = relative_attention_max_distance a_ : Optional[Any] = dropout_rate a_ : Tuple = layer_norm_epsilon a_ : Dict = initializer_factor a_ : Any = feed_forward_proj a_ : Tuple = use_cache a_ : str = add_router_probs a_ : Optional[int] = router_z_loss_coef a_ : List[str] = router_aux_loss_coef a_ : int = self.feed_forward_proj.split('-' ) a_ : int = act_info[-1] a_ : Optional[int] = act_info[0] == 'gated' if len(SCREAMING_SNAKE_CASE__ ) > 1 and act_info[0] != "gated" or len(SCREAMING_SNAKE_CASE__ ) > 2: raise ValueError( F"""`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.""" 'Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. ' '\'gated-gelu\' or \'relu\'' ) # for backwards compatibility if feed_forward_proj == "gated-gelu": a_ : Any = 'gelu_new' super().__init__( pad_token_id=SCREAMING_SNAKE_CASE__ , eos_token_id=SCREAMING_SNAKE_CASE__ , is_encoder_decoder=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ , )
32
0
'''simple docstring''' import argparse import collections import json import os import re import string import sys import numpy as np UpperCamelCase__ : str = re.compile(r'\b(a|an|the)\b', re.UNICODE) UpperCamelCase__ : Any = None def UpperCAmelCase ( ) -> Tuple: """simple docstring""" A_ : Any = argparse.ArgumentParser("""Official evaluation script for SQuAD version 2.0.""" ) parser.add_argument("""data_file""" , metavar="""data.json""" , help="""Input data JSON file.""" ) parser.add_argument("""pred_file""" , metavar="""pred.json""" , help="""Model predictions.""" ) parser.add_argument( """--out-file""" , """-o""" , metavar="""eval.json""" , help="""Write accuracy metrics to file (default is stdout).""" ) parser.add_argument( """--na-prob-file""" , """-n""" , metavar="""na_prob.json""" , help="""Model estimates of probability of no answer.""" ) parser.add_argument( """--na-prob-thresh""" , """-t""" , type=UpperCamelCase__ , default=1.0 , help="""Predict \"\" if no-answer probability exceeds this (default = 1.0).""" , ) parser.add_argument( """--out-image-dir""" , """-p""" , metavar="""out_images""" , default=UpperCamelCase__ , help="""Save precision-recall curves to directory.""" ) parser.add_argument("""--verbose""" , """-v""" , action="""store_true""" ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def UpperCAmelCase ( a_ ) -> Union[str, Any]: """simple docstring""" A_ : Tuple = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: A_ : Dict = bool(qa["""answers"""]["""text"""] ) return qid_to_has_ans def UpperCAmelCase ( a_ ) -> Dict: """simple docstring""" def remove_articles(a_ ): return ARTICLES_REGEX.sub(""" """ , UpperCamelCase__ ) def white_space_fix(a_ ): return " ".join(text.split() ) def remove_punc(a_ ): A_ : Optional[Any] = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(a_ ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(UpperCamelCase__ ) ) ) ) def UpperCAmelCase ( a_ ) -> Any: """simple docstring""" if not s: return [] return normalize_answer(UpperCamelCase__ ).split() def UpperCAmelCase ( a_ , a_ ) -> Any: """simple docstring""" return int(normalize_answer(UpperCamelCase__ ) == normalize_answer(UpperCamelCase__ ) ) def UpperCAmelCase ( a_ , a_ ) -> Dict: """simple docstring""" A_ : Any = get_tokens(UpperCamelCase__ ) A_ : List[Any] = get_tokens(UpperCamelCase__ ) A_ : Optional[int] = collections.Counter(UpperCamelCase__ ) & collections.Counter(UpperCamelCase__ ) A_ : List[str] = sum(common.values() ) if len(UpperCamelCase__ ) == 0 or len(UpperCamelCase__ ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 A_ : Dict = 1.0 * num_same / len(UpperCamelCase__ ) A_ : str = 1.0 * num_same / len(UpperCamelCase__ ) A_ : str = (2 * precision * recall) / (precision + recall) return fa def UpperCAmelCase ( a_ , a_ ) -> Tuple: """simple docstring""" A_ : Optional[int] = {} A_ : List[Any] = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: A_ : str = qa["""id"""] A_ : Tuple = [t for t in qa["""answers"""]["""text"""] if normalize_answer(UpperCamelCase__ )] if not gold_answers: # For unanswerable questions, only correct answer is empty string A_ : Union[str, Any] = [""""""] if qid not in preds: print(F"Missing prediction for {qid}" ) continue A_ : int = preds[qid] # Take max over all gold answers A_ : Optional[int] = max(compute_exact(UpperCamelCase__ , UpperCamelCase__ ) for a in gold_answers ) A_ : int = max(compute_fa(UpperCamelCase__ , UpperCamelCase__ ) for a in gold_answers ) return exact_scores, fa_scores def UpperCAmelCase ( a_ , a_ , a_ , a_ ) -> Optional[int]: """simple docstring""" A_ : List[Any] = {} for qid, s in scores.items(): A_ : Any = na_probs[qid] > na_prob_thresh if pred_na: A_ : Dict = float(not qid_to_has_ans[qid] ) else: A_ : Union[str, Any] = s return new_scores def UpperCAmelCase ( a_ , a_ , a_=None ) -> int: """simple docstring""" if not qid_list: A_ : List[str] = len(UpperCamelCase__ ) return collections.OrderedDict( [ ("""exact""", 100.0 * sum(exact_scores.values() ) / total), ("""f1""", 100.0 * sum(fa_scores.values() ) / total), ("""total""", total), ] ) else: A_ : Tuple = len(UpperCamelCase__ ) return collections.OrderedDict( [ ("""exact""", 100.0 * sum(exact_scores[k] for k in qid_list ) / total), ("""f1""", 100.0 * sum(fa_scores[k] for k in qid_list ) / total), ("""total""", total), ] ) def UpperCAmelCase ( a_ , a_ , a_ ) -> Optional[Any]: """simple docstring""" for k in new_eval: A_ : Optional[int] = new_eval[k] def UpperCAmelCase ( a_ , a_ , a_ , a_ ) -> Tuple: """simple docstring""" plt.step(UpperCamelCase__ , UpperCamelCase__ , color="""b""" , alpha=0.2 , where="""post""" ) plt.fill_between(UpperCamelCase__ , UpperCamelCase__ , step="""post""" , alpha=0.2 , color="""b""" ) plt.xlabel("""Recall""" ) plt.ylabel("""Precision""" ) plt.xlim([0.0, 1.05] ) plt.ylim([0.0, 1.05] ) plt.title(UpperCamelCase__ ) plt.savefig(UpperCamelCase__ ) plt.clf() def UpperCAmelCase ( a_ , a_ , a_ , a_ , a_=None , a_=None ) -> List[Any]: """simple docstring""" A_ : int = sorted(UpperCamelCase__ , key=lambda a_ : na_probs[k] ) A_ : List[str] = 0.0 A_ : Tuple = 1.0 A_ : str = 0.0 A_ : List[Any] = [1.0] A_ : Optional[int] = [0.0] A_ : Any = 0.0 for i, qid in enumerate(UpperCamelCase__ ): if qid_to_has_ans[qid]: true_pos += scores[qid] A_ : Dict = true_pos / float(i + 1 ) A_ : Optional[int] = true_pos / float(UpperCamelCase__ ) if i == len(UpperCamelCase__ ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(UpperCamelCase__ ) recalls.append(UpperCamelCase__ ) if out_image: plot_pr_curve(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) return {"ap": 100.0 * avg_prec} def UpperCAmelCase ( a_ , a_ , a_ , a_ , a_ , a_ ) -> int: """simple docstring""" if out_image_dir and not os.path.exists(UpperCamelCase__ ): os.makedirs(UpperCamelCase__ ) A_ : Optional[int] = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return A_ : Tuple = make_precision_recall_eval( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , out_image=os.path.join(UpperCamelCase__ , """pr_exact.png""" ) , title="""Precision-Recall curve for Exact Match score""" , ) A_ : Tuple = make_precision_recall_eval( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , out_image=os.path.join(UpperCamelCase__ , """pr_f1.png""" ) , title="""Precision-Recall curve for F1 score""" , ) A_ : Any = {k: float(UpperCamelCase__ ) for k, v in qid_to_has_ans.items()} A_ : Any = make_precision_recall_eval( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , out_image=os.path.join(UpperCamelCase__ , """pr_oracle.png""" ) , title="""Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)""" , ) merge_eval(UpperCamelCase__ , UpperCamelCase__ , """pr_exact""" ) merge_eval(UpperCamelCase__ , UpperCamelCase__ , """pr_f1""" ) merge_eval(UpperCamelCase__ , UpperCamelCase__ , """pr_oracle""" ) def UpperCAmelCase ( a_ , a_ , a_ , a_ ) -> Optional[Any]: """simple docstring""" if not qid_list: return A_ : Union[str, Any] = [na_probs[k] for k in qid_list] A_ : List[Any] = np.ones_like(UpperCamelCase__ ) / float(len(UpperCamelCase__ ) ) plt.hist(UpperCamelCase__ , weights=UpperCamelCase__ , bins=2_0 , range=(0.0, 1.0) ) plt.xlabel("""Model probability of no-answer""" ) plt.ylabel("""Proportion of dataset""" ) plt.title(F"Histogram of no-answer probability: {name}" ) plt.savefig(os.path.join(UpperCamelCase__ , F"na_prob_hist_{name}.png" ) ) plt.clf() def UpperCAmelCase ( a_ , a_ , a_ , a_ ) -> str: """simple docstring""" A_ : List[Any] = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) A_ : List[str] = num_no_ans A_ : List[str] = cur_score A_ : List[Any] = 0.0 A_ : Any = sorted(UpperCamelCase__ , key=lambda a_ : na_probs[k] ) for i, qid in enumerate(UpperCamelCase__ ): if qid not in scores: continue if qid_to_has_ans[qid]: A_ : str = scores[qid] else: if preds[qid]: A_ : List[str] = -1 else: A_ : Optional[int] = 0 cur_score += diff if cur_score > best_score: A_ : str = cur_score A_ : Optional[int] = na_probs[qid] return 100.0 * best_score / len(UpperCamelCase__ ), best_thresh def UpperCAmelCase ( a_ , a_ , a_ , a_ , a_ , a_ ) -> List[Any]: """simple docstring""" A_ , A_ : int = find_best_thresh(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) A_ , A_ : Optional[int] = find_best_thresh(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) A_ : int = best_exact A_ : int = exact_thresh A_ : str = best_fa A_ : str = fa_thresh def UpperCAmelCase ( ) -> str: """simple docstring""" with open(OPTS.data_file ) as f: A_ : Any = json.load(UpperCamelCase__ ) A_ : str = dataset_json["""data"""] with open(OPTS.pred_file ) as f: A_ : Tuple = json.load(UpperCamelCase__ ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: A_ : Optional[int] = json.load(UpperCamelCase__ ) else: A_ : Optional[int] = {k: 0.0 for k in preds} A_ : Optional[int] = make_qid_to_has_ans(UpperCamelCase__ ) # maps qid to True/False A_ : int = [k for k, v in qid_to_has_ans.items() if v] A_ : List[str] = [k for k, v in qid_to_has_ans.items() if not v] A_ , A_ : Optional[int] = get_raw_scores(UpperCamelCase__ , UpperCamelCase__ ) A_ : Optional[int] = apply_no_ans_threshold(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , OPTS.na_prob_thresh ) A_ : str = apply_no_ans_threshold(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , OPTS.na_prob_thresh ) A_ : List[Any] = make_eval_dict(UpperCamelCase__ , UpperCamelCase__ ) if has_ans_qids: A_ : str = make_eval_dict(UpperCamelCase__ , UpperCamelCase__ , qid_list=UpperCamelCase__ ) merge_eval(UpperCamelCase__ , UpperCamelCase__ , """HasAns""" ) if no_ans_qids: A_ : Optional[Any] = make_eval_dict(UpperCamelCase__ , UpperCamelCase__ , qid_list=UpperCamelCase__ ) merge_eval(UpperCamelCase__ , UpperCamelCase__ , """NoAns""" ) if OPTS.na_prob_file: find_all_best_thresh(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , OPTS.out_image_dir ) histogram_na_prob(UpperCamelCase__ , UpperCamelCase__ , OPTS.out_image_dir , """hasAns""" ) histogram_na_prob(UpperCamelCase__ , UpperCamelCase__ , OPTS.out_image_dir , """noAns""" ) if OPTS.out_file: with open(OPTS.out_file , """w""" ) as f: json.dump(UpperCamelCase__ , UpperCamelCase__ ) else: print(json.dumps(UpperCamelCase__ , indent=2 ) ) if __name__ == "__main__": UpperCamelCase__ : Union[str, Any] = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use('Agg') import matplotlib.pyplot as plt main()
363
'''simple docstring''' import unittest from transformers import MPNetConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=13 , _lowerCamelCase=7 , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=False , _lowerCamelCase=True , _lowerCamelCase=99 , _lowerCamelCase=64 , _lowerCamelCase=5 , _lowerCamelCase=4 , _lowerCamelCase=64 , _lowerCamelCase="gelu" , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=512 , _lowerCamelCase=16 , _lowerCamelCase=2 , _lowerCamelCase=0.02 , _lowerCamelCase=3 , _lowerCamelCase=4 , _lowerCamelCase=None , ) -> Union[str, Any]: A_ : Tuple = parent A_ : Optional[Any] = batch_size A_ : Optional[Any] = seq_length A_ : List[str] = is_training A_ : str = use_input_mask A_ : List[str] = use_token_type_ids A_ : Tuple = use_labels A_ : List[str] = vocab_size A_ : Optional[Any] = hidden_size A_ : Any = num_hidden_layers A_ : Optional[int] = num_attention_heads A_ : Optional[int] = intermediate_size A_ : str = hidden_act A_ : Union[str, Any] = hidden_dropout_prob A_ : int = attention_probs_dropout_prob A_ : Dict = max_position_embeddings A_ : Dict = type_vocab_size A_ : Any = type_sequence_label_size A_ : Optional[int] = initializer_range A_ : int = num_labels A_ : int = num_choices A_ : Optional[int] = scope def UpperCAmelCase_ ( self ) -> Optional[int]: return MPNetConfig.from_pretrained("""microsoft/mpnet-base""" ) def UpperCAmelCase_ ( self ) -> Tuple: A_ : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A_ : Dict = None if self.use_input_mask: A_ : int = random_attention_mask([self.batch_size, self.seq_length] ) A_ : Tuple = None A_ : Optional[int] = None A_ : Union[str, Any] = None if self.use_labels: A_ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A_ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A_ : Optional[int] = ids_tensor([self.batch_size] , self.num_choices ) A_ : int = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase_ ( self ) -> int: return MPNetConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> List[str]: A_ : int = MPNetModel(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : Union[str, Any] = model(_lowerCamelCase , _lowerCamelCase ) A_ : Optional[Any] = model(_lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Union[str, Any]: A_ : str = MPNetForQuestionAnswering(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : Optional[int] = model( _lowerCamelCase , attention_mask=_lowerCamelCase , start_positions=_lowerCamelCase , end_positions=_lowerCamelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> List[str]: A_ : Tuple = self.num_labels A_ : List[Any] = MPNetForSequenceClassification(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : str = model(_lowerCamelCase , attention_mask=_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Dict: A_ : int = self.num_choices A_ : Dict = MPNetForMultipleChoice(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : Union[str, Any] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A_ : Dict = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A_ : str = model( _lowerCamelCase , attention_mask=_lowerCamelCase , labels=_lowerCamelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Optional[Any]: A_ : Optional[int] = self.num_labels A_ : Tuple = MPNetForTokenClassification(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() A_ : Union[str, Any] = model(_lowerCamelCase , attention_mask=_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCAmelCase_ ( self ) -> str: A_ : int = self.prepare_config_and_inputs() ((A_) , (A_) , (A_) , (A_) , (A_) , (A_)) : Any = config_and_inputs A_ : Tuple = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class _lowerCAmelCase ( __A, __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = ( ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) if is_torch_available() else () ) lowerCamelCase = ( { '''feature-extraction''': MPNetModel, '''fill-mask''': MPNetForMaskedLM, '''question-answering''': MPNetForQuestionAnswering, '''text-classification''': MPNetForSequenceClassification, '''token-classification''': MPNetForTokenClassification, '''zero-shot''': MPNetForSequenceClassification, } if is_torch_available() else {} ) lowerCamelCase = False lowerCamelCase = True def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Dict = MPNetModelTester(self ) A_ : int = ConfigTester(self , config_class=_lowerCamelCase , hidden_size=37 ) def UpperCAmelCase_ ( self ) -> int: self.config_tester.run_common_tests() def UpperCAmelCase_ ( self ) -> Any: A_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_model(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_sequence_classification(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> int: A_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_multiple_choice(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> List[str]: A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_token_classification(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_question_answering(*_lowerCamelCase ) @require_torch class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def UpperCAmelCase_ ( self ) -> Tuple: A_ : int = MPNetModel.from_pretrained("""microsoft/mpnet-base""" ) A_ : int = torch.tensor([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) A_ : Tuple = model(_lowerCamelCase )[0] A_ : int = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , _lowerCamelCase ) A_ : Any = torch.tensor( [[[-0.0550, 0.1943, -0.0740], [-0.0562, 0.2211, -0.0579], [-0.0437, 0.3337, -0.0641]]] ) # compare the actual values for a slice. self.assertTrue(torch.allclose(output[:, :3, :3] , _lowerCamelCase , atol=1e-4 ) )
164
0
import operator as op def lowerCamelCase__ ( A__ : Optional[int] ): '''simple docstring''' __lowerCamelCase = [] __lowerCamelCase = lambda A__ , A__ : int(x / y ) # noqa: E731 integer division operation __lowerCamelCase = { """^""": op.pow, """*""": op.mul, """/""": div, """+""": op.add, """-""": op.sub, } # operators & their respective operation # print table header print("""Symbol""".center(8 ) , """Action""".center(12 ) , """Stack""" , sep=""" | """ ) print("""-""" * (30 + len(_lowercase )) ) for x in post_fix: if x.isdigit(): # if x in digit stack.append(_lowercase ) # append x to stack # output in tabular format print(x.rjust(8 ) , ("""push(""" + x + """)""").ljust(12 ) , """,""".join(_lowercase ) , sep=""" | """ ) else: __lowerCamelCase = stack.pop() # pop stack # output in tabular format print("""""".rjust(8 ) , ("""pop(""" + b + """)""").ljust(12 ) , """,""".join(_lowercase ) , sep=""" | """ ) __lowerCamelCase = stack.pop() # pop stack # output in tabular format print("""""".rjust(8 ) , ("""pop(""" + a + """)""").ljust(12 ) , """,""".join(_lowercase ) , sep=""" | """ ) stack.append( str(opr[x](int(_lowercase ) , int(_lowercase ) ) ) ) # evaluate the 2 values popped from stack & push result to stack # output in tabular format print( x.rjust(8 ) , ("""push(""" + a + x + b + """)""").ljust(12 ) , """,""".join(_lowercase ) , sep=""" | """ , ) return int(stack[0] ) if __name__ == "__main__": UpperCAmelCase_ = input('\n\nEnter a Postfix Equation (space separated) = ').split(' ') print('\n\tResult = ', solve(Postfix))
12
"""simple docstring""" import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": __a = pd.read_csv("sample_data.csv", header=None) __a = df.shape[:1][0] # If you're using some other dataset input the target column __a = df.iloc[:, 1:2] __a = actual_data.values.reshape(len_data, 1) __a = MinMaxScaler().fit_transform(actual_data) __a = 10 __a = 5 __a = 20 __a = len_data - periods * look_back __a = actual_data[:division] __a = actual_data[division - look_back :] __a , __a = [], [] __a , __a = [], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) __a = np.array(train_x) __a = np.array(test_x) __a = np.array([list(i.ravel()) for i in train_y]) __a = np.array([list(i.ravel()) for i in test_y]) __a = Sequential() model.add(LSTM(1_28, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(64, input_shape=(1_28, 1))) model.add(Dense(forward_days)) model.compile(loss="mean_squared_error", optimizer="adam") __a = model.fit( x_train, y_train, epochs=1_50, verbose=1, shuffle=True, batch_size=4 ) __a = model.predict(x_test)
66
0
"""simple docstring""" lowercase__ = '\n# Transformers 설치 방법\n! pip install transformers datasets\n# 마지막 릴리스 대신 소스에서 설치하려면, 위 명령을 주석으로 바꾸고 아래 명령을 해제하세요.\n# ! pip install git+https://github.com/huggingface/transformers.git\n' lowercase__ = [{'type': 'code', 'content': INSTALL_CONTENT}] lowercase__ = { '{processor_class}': 'FakeProcessorClass', '{model_class}': 'FakeModelClass', '{object_class}': 'FakeObjectClass', }
353
"""simple docstring""" import argparse import collections import json import os import re import string import sys import numpy as np lowercase__ = re.compile(r'\b(a|an|the)\b', re.UNICODE) lowercase__ = None def __a ( ) ->List[Any]: a__: Dict = argparse.ArgumentParser('Official evaluation script for SQuAD version 2.0.' ) parser.add_argument('data_file' , metavar='data.json' , help='Input data JSON file.' ) parser.add_argument('pred_file' , metavar='pred.json' , help='Model predictions.' ) parser.add_argument( '--out-file' , '-o' , metavar='eval.json' , help='Write accuracy metrics to file (default is stdout).' ) parser.add_argument( '--na-prob-file' , '-n' , metavar='na_prob.json' , help='Model estimates of probability of no answer.' ) parser.add_argument( '--na-prob-thresh' , '-t' , type=_SCREAMING_SNAKE_CASE , default=1.0 , help='Predict "" if no-answer probability exceeds this (default = 1.0).' , ) parser.add_argument( '--out-image-dir' , '-p' , metavar='out_images' , default=_SCREAMING_SNAKE_CASE , help='Save precision-recall curves to directory.' ) parser.add_argument('--verbose' , '-v' , action='store_true' ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def __a ( _SCREAMING_SNAKE_CASE ) ->Union[str, Any]: a__: Optional[int] = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: a__: Optional[Any] = bool(qa['answers']['text'] ) return qid_to_has_ans def __a ( _SCREAMING_SNAKE_CASE ) ->Optional[Any]: def remove_articles(_SCREAMING_SNAKE_CASE ): return ARTICLES_REGEX.sub(' ' , _SCREAMING_SNAKE_CASE ) def white_space_fix(_SCREAMING_SNAKE_CASE ): return " ".join(text.split() ) def remove_punc(_SCREAMING_SNAKE_CASE ): a__: Dict = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(_SCREAMING_SNAKE_CASE ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(_SCREAMING_SNAKE_CASE ) ) ) ) def __a ( _SCREAMING_SNAKE_CASE ) ->Optional[int]: if not s: return [] return normalize_answer(_SCREAMING_SNAKE_CASE ).split() def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->List[Any]: return int(normalize_answer(_SCREAMING_SNAKE_CASE ) == normalize_answer(_SCREAMING_SNAKE_CASE ) ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->str: a__: Any = get_tokens(_SCREAMING_SNAKE_CASE ) a__: Optional[int] = get_tokens(_SCREAMING_SNAKE_CASE ) a__: Optional[int] = collections.Counter(_SCREAMING_SNAKE_CASE ) & collections.Counter(_SCREAMING_SNAKE_CASE ) a__: Tuple = sum(common.values() ) if len(_SCREAMING_SNAKE_CASE ) == 0 or len(_SCREAMING_SNAKE_CASE ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 a__: Any = 1.0 * num_same / len(_SCREAMING_SNAKE_CASE ) a__: Optional[int] = 1.0 * num_same / len(_SCREAMING_SNAKE_CASE ) a__: Dict = (2 * precision * recall) / (precision + recall) return fa def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Dict: a__: Union[str, Any] = {} a__: Dict = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: a__: Optional[int] = qa['id'] a__: List[Any] = [t for t in qa['answers']['text'] if normalize_answer(_SCREAMING_SNAKE_CASE )] if not gold_answers: # For unanswerable questions, only correct answer is empty string a__: str = [''] if qid not in preds: print(F'Missing prediction for {qid}' ) continue a__: Any = preds[qid] # Take max over all gold answers a__: List[str] = max(compute_exact(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for a in gold_answers ) a__: Optional[int] = max(compute_fa(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for a in gold_answers ) return exact_scores, fa_scores def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[Any]: a__: List[str] = {} for qid, s in scores.items(): a__: List[Any] = na_probs[qid] > na_prob_thresh if pred_na: a__: Optional[int] = float(not qid_to_has_ans[qid] ) else: a__: Optional[Any] = s return new_scores def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None ) ->Tuple: if not qid_list: a__: str = len(_SCREAMING_SNAKE_CASE ) return collections.OrderedDict( [ ('exact', 100.0 * sum(exact_scores.values() ) / total), ('f1', 100.0 * sum(fa_scores.values() ) / total), ('total', total), ] ) else: a__: Optional[Any] = len(_SCREAMING_SNAKE_CASE ) return collections.OrderedDict( [ ('exact', 100.0 * sum(exact_scores[k] for k in qid_list ) / total), ('f1', 100.0 * sum(fa_scores[k] for k in qid_list ) / total), ('total', total), ] ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->List[str]: for k in new_eval: a__: List[Any] = new_eval[k] def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->str: plt.step(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , color='b' , alpha=0.2 , where='post' ) plt.fill_between(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , step='post' , alpha=0.2 , color='b' ) plt.xlabel('Recall' ) plt.ylabel('Precision' ) plt.xlim([0.0, 1.05] ) plt.ylim([0.0, 1.05] ) plt.title(_SCREAMING_SNAKE_CASE ) plt.savefig(_SCREAMING_SNAKE_CASE ) plt.clf() def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None ) ->List[str]: a__: Optional[int] = sorted(_SCREAMING_SNAKE_CASE , key=lambda _SCREAMING_SNAKE_CASE : na_probs[k] ) a__: Dict = 0.0 a__: Optional[int] = 1.0 a__: Tuple = 0.0 a__: Tuple = [1.0] a__: Optional[Any] = [0.0] a__: Optional[Any] = 0.0 for i, qid in enumerate(_SCREAMING_SNAKE_CASE ): if qid_to_has_ans[qid]: true_pos += scores[qid] a__: Optional[Any] = true_pos / float(i + 1 ) a__: int = true_pos / float(_SCREAMING_SNAKE_CASE ) if i == len(_SCREAMING_SNAKE_CASE ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(_SCREAMING_SNAKE_CASE ) recalls.append(_SCREAMING_SNAKE_CASE ) if out_image: plot_pr_curve(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) return {"ap": 100.0 * avg_prec} def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: if out_image_dir and not os.path.exists(_SCREAMING_SNAKE_CASE ): os.makedirs(_SCREAMING_SNAKE_CASE ) a__: Any = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return a__: Optional[Any] = make_precision_recall_eval( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , out_image=os.path.join(_SCREAMING_SNAKE_CASE , 'pr_exact.png' ) , title='Precision-Recall curve for Exact Match score' , ) a__: List[str] = make_precision_recall_eval( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , out_image=os.path.join(_SCREAMING_SNAKE_CASE , 'pr_f1.png' ) , title='Precision-Recall curve for F1 score' , ) a__: Optional[Any] = {k: float(_SCREAMING_SNAKE_CASE ) for k, v in qid_to_has_ans.items()} a__: List[Any] = make_precision_recall_eval( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , out_image=os.path.join(_SCREAMING_SNAKE_CASE , 'pr_oracle.png' ) , title='Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)' , ) merge_eval(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , 'pr_exact' ) merge_eval(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , 'pr_f1' ) merge_eval(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , 'pr_oracle' ) def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->List[str]: if not qid_list: return a__: Any = [na_probs[k] for k in qid_list] a__: List[str] = np.ones_like(_SCREAMING_SNAKE_CASE ) / float(len(_SCREAMING_SNAKE_CASE ) ) plt.hist(_SCREAMING_SNAKE_CASE , weights=_SCREAMING_SNAKE_CASE , bins=20 , range=(0.0, 1.0) ) plt.xlabel('Model probability of no-answer' ) plt.ylabel('Proportion of dataset' ) plt.title(F'Histogram of no-answer probability: {name}' ) plt.savefig(os.path.join(_SCREAMING_SNAKE_CASE , F'na_prob_hist_{name}.png' ) ) plt.clf() def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->List[str]: a__: str = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) a__: List[Any] = num_no_ans a__: Union[str, Any] = cur_score a__: Optional[Any] = 0.0 a__: str = sorted(_SCREAMING_SNAKE_CASE , key=lambda _SCREAMING_SNAKE_CASE : na_probs[k] ) for i, qid in enumerate(_SCREAMING_SNAKE_CASE ): if qid not in scores: continue if qid_to_has_ans[qid]: a__: Tuple = scores[qid] else: if preds[qid]: a__: Optional[Any] = -1 else: a__: Optional[int] = 0 cur_score += diff if cur_score > best_score: a__: Dict = cur_score a__: Optional[int] = na_probs[qid] return 100.0 * best_score / len(_SCREAMING_SNAKE_CASE ), best_thresh def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->List[str]: a__ , a__: str = find_best_thresh(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__ , a__: Optional[int] = find_best_thresh(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: List[Any] = best_exact a__: Dict = exact_thresh a__: Optional[int] = best_fa a__: str = fa_thresh def __a ( ) ->int: with open(OPTS.data_file ) as f: a__: Tuple = json.load(_SCREAMING_SNAKE_CASE ) a__: Union[str, Any] = dataset_json['data'] with open(OPTS.pred_file ) as f: a__: Dict = json.load(_SCREAMING_SNAKE_CASE ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: a__: Dict = json.load(_SCREAMING_SNAKE_CASE ) else: a__: Optional[Any] = {k: 0.0 for k in preds} a__: List[Any] = make_qid_to_has_ans(_SCREAMING_SNAKE_CASE ) # maps qid to True/False a__: Optional[int] = [k for k, v in qid_to_has_ans.items() if v] a__: Union[str, Any] = [k for k, v in qid_to_has_ans.items() if not v] a__ , a__: Optional[Any] = get_raw_scores(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) a__: Any = apply_no_ans_threshold(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , OPTS.na_prob_thresh ) a__: Dict = apply_no_ans_threshold(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , OPTS.na_prob_thresh ) a__: str = make_eval_dict(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if has_ans_qids: a__: List[str] = make_eval_dict(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , qid_list=_SCREAMING_SNAKE_CASE ) merge_eval(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , 'HasAns' ) if no_ans_qids: a__: Optional[Any] = make_eval_dict(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , qid_list=_SCREAMING_SNAKE_CASE ) merge_eval(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , 'NoAns' ) if OPTS.na_prob_file: find_all_best_thresh(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , OPTS.out_image_dir ) histogram_na_prob(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , OPTS.out_image_dir , 'hasAns' ) histogram_na_prob(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , OPTS.out_image_dir , 'noAns' ) if OPTS.out_file: with open(OPTS.out_file , 'w' ) as f: json.dump(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) else: print(json.dumps(_SCREAMING_SNAKE_CASE , indent=2 ) ) if __name__ == "__main__": lowercase__ = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use('Agg') import matplotlib.pyplot as plt main()
203
0
"""simple docstring""" import os from collections import namedtuple import pytest from datasets import ClassLabel, Features, Sequence, Value from datasets.commands.test import TestCommand from datasets.info import DatasetInfo, DatasetInfosDict _a : Optional[int]= namedtuple( "_TestCommandArgs", [ "dataset", "name", "cache_dir", "data_dir", "all_configs", "save_infos", "ignore_verifications", "force_redownload", "clear_cache", ], defaults=[None, None, None, False, False, False, False, False], ) def __UpperCAmelCase ( UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Tuple ) -> List[Any]: '''simple docstring''' return (abs(source - target ) / target) < 0.01 @pytest.mark.integration def __UpperCAmelCase ( UpperCAmelCase_ : Optional[int] ) -> Optional[Any]: '''simple docstring''' __snake_case : Dict = _TestCommandArgs(dataset=lowerCAmelCase__ , all_configs=lowerCAmelCase__ , save_infos=lowerCAmelCase__ ) __snake_case : List[str] = TestCommand(*lowerCAmelCase__ ) test_command.run() __snake_case : List[str] = os.path.join(lowerCAmelCase__ , 'README.md' ) assert os.path.exists(lowerCAmelCase__ ) __snake_case : Dict = DatasetInfosDict.from_directory(lowerCAmelCase__ ) __snake_case : Optional[Any] = DatasetInfosDict( { 'default': DatasetInfo( features=Features( { 'tokens': Sequence(Value('string' ) ), 'ner_tags': Sequence( ClassLabel(names=['O', 'B-PER', 'I-PER', 'B-ORG', 'I-ORG', 'B-LOC', 'I-LOC'] ) ), 'langs': Sequence(Value('string' ) ), 'spans': Sequence(Value('string' ) ), } ) , splits=[ { 'name': 'train', 'num_bytes': 2_35_15_63, 'num_examples': 1_00_00, }, { 'name': 'validation', 'num_bytes': 23_84_18, 'num_examples': 10_00, }, ] , download_size=3_94_06_80 , dataset_size=2_58_99_81 , ) } ) assert dataset_infos.keys() == expected_dataset_infos.keys() for key in DatasetInfo._INCLUDED_INFO_IN_YAML: __snake_case : Any = getattr(dataset_infos['default'] , lowerCAmelCase__ ), getattr(expected_dataset_infos['default'] , lowerCAmelCase__ ) if key == "num_bytes": assert is_apercent_close(lowerCAmelCase__ , lowerCAmelCase__ ) elif key == "splits": assert list(lowerCAmelCase__ ) == list(lowerCAmelCase__ ) for split in result: assert result[split].name == expected[split].name assert result[split].num_examples == expected[split].num_examples assert is_apercent_close(result[split].num_bytes , expected[split].num_bytes ) else: result == expected
172
import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTokenizerFast from utils import get_device, get_timestamp, show_pil class _a : def __init__(self, SCREAMING_SNAKE_CASE_ = "cpu", SCREAMING_SNAKE_CASE_ = "openai/clip-vit-large-patch14" ) -> None: UpperCAmelCase_: Optional[Any] = device UpperCAmelCase_: Optional[Any] = CLIPTokenizerFast.from_pretrained(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Tuple = [0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3] UpperCAmelCase_: Optional[Any] = [0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1] UpperCAmelCase_: Optional[Any] = torchvision.transforms.Normalize(self.image_mean, self.image_std ) UpperCAmelCase_: Tuple = torchvision.transforms.Resize(224 ) UpperCAmelCase_: Any = torchvision.transforms.CenterCrop(224 ) def __snake_case (self, SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: UpperCAmelCase_: Dict = self.resize(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Optional[int] = self.center_crop(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Tuple = self.normalize(SCREAMING_SNAKE_CASE_ ) return images def __call__(self, SCREAMING_SNAKE_CASE_=None, SCREAMING_SNAKE_CASE_=None, **SCREAMING_SNAKE_CASE_ ) -> List[str]: UpperCAmelCase_: Dict = self.tokenizer(text=SCREAMING_SNAKE_CASE_, **SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Tuple = self.preprocess_img(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: int = {key: value.to(self.device ) for (key, value) in encoding.items()} return encoding class _a ( nn.Module ): def __init__(self, SCREAMING_SNAKE_CASE_=10, SCREAMING_SNAKE_CASE_=0.0_1, SCREAMING_SNAKE_CASE_=None, SCREAMING_SNAKE_CASE_=None, SCREAMING_SNAKE_CASE_=None, SCREAMING_SNAKE_CASE_=None, SCREAMING_SNAKE_CASE_=None, SCREAMING_SNAKE_CASE_=None, SCREAMING_SNAKE_CASE_=False, SCREAMING_SNAKE_CASE_=True, SCREAMING_SNAKE_CASE_="image", SCREAMING_SNAKE_CASE_=True, SCREAMING_SNAKE_CASE_=False, SCREAMING_SNAKE_CASE_=False, SCREAMING_SNAKE_CASE_=False, ) -> None: super().__init__() UpperCAmelCase_: List[Any] = None UpperCAmelCase_: List[str] = device if device else get_device() if vqgan: UpperCAmelCase_: int = vqgan else: UpperCAmelCase_: Optional[Any] = load_vqgan(self.device, conf_path=SCREAMING_SNAKE_CASE_, ckpt_path=SCREAMING_SNAKE_CASE_ ) self.vqgan.eval() if clip: UpperCAmelCase_: List[str] = clip else: UpperCAmelCase_: Any = CLIPModel.from_pretrained("""openai/clip-vit-base-patch32""" ) self.clip.to(self.device ) UpperCAmelCase_: Optional[int] = ProcessorGradientFlow(device=self.device ) UpperCAmelCase_: Optional[int] = iterations UpperCAmelCase_: List[Any] = lr UpperCAmelCase_: str = log UpperCAmelCase_: Tuple = make_grid UpperCAmelCase_: List[str] = return_val UpperCAmelCase_: Dict = quantize UpperCAmelCase_: int = self.vqgan.decoder.z_shape def __snake_case (self, SCREAMING_SNAKE_CASE_=None, SCREAMING_SNAKE_CASE_=None, SCREAMING_SNAKE_CASE_=5, SCREAMING_SNAKE_CASE_=True ) -> List[Any]: UpperCAmelCase_: Tuple = [] if output_path is None: UpperCAmelCase_: Optional[int] = """./animation.gif""" if input_path is None: UpperCAmelCase_: Tuple = self.save_path UpperCAmelCase_: List[Any] = sorted(glob(input_path + """/*""" ) ) if not len(SCREAMING_SNAKE_CASE_ ): raise ValueError( """No images found in save path, aborting (did you pass save_intermediate=True to the generate""" """ function?)""" ) if len(SCREAMING_SNAKE_CASE_ ) == 1: print("""Only one image found in save path, (did you pass save_intermediate=True to the generate function?)""" ) UpperCAmelCase_: Dict = total_duration / len(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: str = [frame_duration] * len(SCREAMING_SNAKE_CASE_ ) if extend_frames: UpperCAmelCase_: List[str] = 1.5 UpperCAmelCase_: List[Any] = 3 for file_name in paths: if file_name.endswith(""".png""" ): images.append(imageio.imread(SCREAMING_SNAKE_CASE_ ) ) imageio.mimsave(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, duration=SCREAMING_SNAKE_CASE_ ) print(f'gif saved to {output_path}' ) def __snake_case (self, SCREAMING_SNAKE_CASE_=None, SCREAMING_SNAKE_CASE_=None ) -> Optional[int]: if not (path or img): raise ValueError("""Input either path or tensor""" ) if img is not None: raise NotImplementedError UpperCAmelCase_: List[Any] = preprocess(Image.open(SCREAMING_SNAKE_CASE_ ), target_image_size=256 ).to(self.device ) UpperCAmelCase_: Union[str, Any] = preprocess_vqgan(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ , *UpperCAmelCase_: str = self.vqgan.encode(SCREAMING_SNAKE_CASE_ ) return z def __snake_case (self, SCREAMING_SNAKE_CASE_ ) -> Union[str, Any]: UpperCAmelCase_: List[Any] = self.latent.detach().requires_grad_() UpperCAmelCase_: Optional[int] = base_latent + transform_vector if self.quantize: UpperCAmelCase_ , *UpperCAmelCase_: Optional[Any] = self.vqgan.quantize(SCREAMING_SNAKE_CASE_ ) else: UpperCAmelCase_: Tuple = trans_latent return self.vqgan.decode(SCREAMING_SNAKE_CASE_ ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_=None ) -> List[str]: UpperCAmelCase_: Any = self.clip_preprocessor(text=SCREAMING_SNAKE_CASE_, images=SCREAMING_SNAKE_CASE_, return_tensors="""pt""", padding=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: str = self.clip(**SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Dict = clip_outputs.logits_per_image if weights is not None: UpperCAmelCase_: Any = similarity_logits * weights return similarity_logits.sum() def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Any: UpperCAmelCase_: Dict = self._get_clip_similarity(pos_prompts["""prompts"""], SCREAMING_SNAKE_CASE_, weights=(1 / pos_prompts["""weights"""]) ) if neg_prompts: UpperCAmelCase_: Tuple = self._get_clip_similarity(neg_prompts["""prompts"""], SCREAMING_SNAKE_CASE_, weights=neg_prompts["""weights"""] ) else: UpperCAmelCase_: Any = torch.tensor([1], device=self.device ) UpperCAmelCase_: List[str] = -torch.log(SCREAMING_SNAKE_CASE_ ) + torch.log(SCREAMING_SNAKE_CASE_ ) return loss def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: UpperCAmelCase_: Tuple = torch.randn_like(self.latent, requires_grad=SCREAMING_SNAKE_CASE_, device=self.device ) UpperCAmelCase_: str = torch.optim.Adam([vector], lr=self.lr ) for i in range(self.iterations ): optim.zero_grad() UpperCAmelCase_: Optional[int] = self._add_vector(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Any = loop_post_process(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: int = self._get_CLIP_loss(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) print("""CLIP loss""", SCREAMING_SNAKE_CASE_ ) if self.log: wandb.log({"""CLIP Loss""": clip_loss} ) clip_loss.backward(retain_graph=SCREAMING_SNAKE_CASE_ ) optim.step() if self.return_val == "image": yield custom_to_pil(transformed_img[0] ) else: yield vector def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Dict: wandb.init(reinit=SCREAMING_SNAKE_CASE_, project="""face-editor""" ) wandb.config.update({"""Positive Prompts""": positive_prompts} ) wandb.config.update({"""Negative Prompts""": negative_prompts} ) wandb.config.update({"""lr""": self.lr, """iterations""": self.iterations} ) if image_path: UpperCAmelCase_: str = Image.open(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Optional[Any] = image.resize((256, 256) ) wandb.log("""Original Image""", wandb.Image(SCREAMING_SNAKE_CASE_ ) ) def __snake_case (self, SCREAMING_SNAKE_CASE_ ) -> Optional[int]: if not prompts: return [] UpperCAmelCase_: Tuple = [] UpperCAmelCase_: str = [] if isinstance(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ): UpperCAmelCase_: Optional[Any] = [prompt.strip() for prompt in prompts.split("""|""" )] for prompt in prompts: if isinstance(SCREAMING_SNAKE_CASE_, (tuple, list) ): UpperCAmelCase_: str = prompt[0] UpperCAmelCase_: List[str] = float(prompt[1] ) elif ":" in prompt: UpperCAmelCase_ , UpperCAmelCase_: int = prompt.split(""":""" ) UpperCAmelCase_: int = float(SCREAMING_SNAKE_CASE_ ) else: UpperCAmelCase_: str = prompt UpperCAmelCase_: Dict = 1.0 processed_prompts.append(SCREAMING_SNAKE_CASE_ ) weights.append(SCREAMING_SNAKE_CASE_ ) return { "prompts": processed_prompts, "weights": torch.tensor(SCREAMING_SNAKE_CASE_, device=self.device ), } def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_=None, SCREAMING_SNAKE_CASE_=None, SCREAMING_SNAKE_CASE_=True, SCREAMING_SNAKE_CASE_=False, SCREAMING_SNAKE_CASE_=True, SCREAMING_SNAKE_CASE_=True, SCREAMING_SNAKE_CASE_=None, ) -> Optional[Any]: if image_path: UpperCAmelCase_: Optional[int] = self._get_latent(SCREAMING_SNAKE_CASE_ ) else: UpperCAmelCase_: str = torch.randn(self.latent_dim, device=self.device ) if self.log: self._init_logging(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) assert pos_prompts, "You must provide at least one positive prompt." UpperCAmelCase_: List[Any] = self.process_prompts(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Dict = self.process_prompts(SCREAMING_SNAKE_CASE_ ) if save_final and save_path is None: UpperCAmelCase_: Optional[int] = os.path.join("""./outputs/""", """_""".join(pos_prompts["""prompts"""] ) ) if not os.path.exists(SCREAMING_SNAKE_CASE_ ): os.makedirs(SCREAMING_SNAKE_CASE_ ) else: UpperCAmelCase_: List[str] = save_path + """_""" + get_timestamp() os.makedirs(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Optional[int] = save_path UpperCAmelCase_: Optional[Any] = self.vqgan.decode(self.latent )[0] if show_intermediate: print("""Original Image""" ) show_pil(custom_to_pil(SCREAMING_SNAKE_CASE_ ) ) UpperCAmelCase_: Tuple = loop_post_process(SCREAMING_SNAKE_CASE_ ) for iter, transformed_img in enumerate(self._optimize_CLIP(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) ): if show_intermediate: show_pil(SCREAMING_SNAKE_CASE_ ) if save_intermediate: transformed_img.save(os.path.join(self.save_path, f'iter_{iter:03d}.png' ) ) if self.log: wandb.log({"""Image""": wandb.Image(SCREAMING_SNAKE_CASE_ )} ) if show_final: show_pil(SCREAMING_SNAKE_CASE_ ) if save_final: transformed_img.save(os.path.join(self.save_path, f'iter_{iter:03d}_final.png' ) )
147
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCAmelCase : Optional[Any] = { "configuration_roc_bert": ["ROC_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "RoCBertConfig"], "tokenization_roc_bert": ["RoCBertTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: pass try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : int = [ "ROC_BERT_PRETRAINED_MODEL_ARCHIVE_LIST", "RoCBertForCausalLM", "RoCBertForMaskedLM", "RoCBertForMultipleChoice", "RoCBertForPreTraining", "RoCBertForQuestionAnswering", "RoCBertForSequenceClassification", "RoCBertForTokenClassification", "RoCBertLayer", "RoCBertModel", "RoCBertPreTrainedModel", "load_tf_weights_in_roc_bert", ] if TYPE_CHECKING: from .configuration_roc_bert import ROC_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, RoCBertConfig from .tokenization_roc_bert import RoCBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: raise OptionalDependencyNotAvailable() try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roc_bert import ( ROC_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, RoCBertForCausalLM, RoCBertForMaskedLM, RoCBertForMultipleChoice, RoCBertForPreTraining, RoCBertForQuestionAnswering, RoCBertForSequenceClassification, RoCBertForTokenClassification, RoCBertLayer, RoCBertModel, RoCBertPreTrainedModel, load_tf_weights_in_roc_bert, ) else: import sys UpperCAmelCase : int = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
368
'''simple docstring''' def a__ ( a__ ): """simple docstring""" __SCREAMING_SNAKE_CASE = len(a__ ) while cur > 1: # Find the maximum number in arr __SCREAMING_SNAKE_CASE = arr.index(max(arr[0:cur] ) ) # Reverse from 0 to mi __SCREAMING_SNAKE_CASE = arr[mi::-1] + arr[mi + 1 : len(a__ )] # Reverse whole list __SCREAMING_SNAKE_CASE = arr[cur - 1 :: -1] + arr[cur : len(a__ )] cur -= 1 return arr if __name__ == "__main__": UpperCAmelCase : Tuple = input('Enter numbers separated by a comma:\n').strip() UpperCAmelCase : str = [int(item) for item in user_input.split(',')] print(pancake_sort(unsorted))
331
0
'''simple docstring''' import random def _UpperCamelCase ( __A , __A , __A ) -> List[Any]: '''simple docstring''' UpperCamelCase__ = a[left_index] UpperCamelCase__ = left_index + 1 for j in range(left_index + 1 , __A ): if a[j] < pivot: UpperCamelCase__ , UpperCamelCase__ = a[i], a[j] i += 1 UpperCamelCase__ , UpperCamelCase__ = a[i - 1], a[left_index] return i - 1 def _UpperCamelCase ( __A , __A , __A ) -> Any: '''simple docstring''' if left < right: UpperCamelCase__ = random.randint(__A , right - 1 ) UpperCamelCase__ , UpperCamelCase__ = ( a[left], a[pivot], ) # switches the pivot with the left most bound UpperCamelCase__ = partition(__A , __A , __A ) quick_sort_random( __A , __A , __A ) # recursive quicksort to the left of the pivot point quick_sort_random( __A , pivot_index + 1 , __A ) # recursive quicksort to the right of the pivot point def _UpperCamelCase ( ) -> int: '''simple docstring''' UpperCamelCase__ = input("Enter numbers separated by a comma:\n" ).strip() UpperCamelCase__ = [int(__A ) for item in user_input.split("," )] quick_sort_random(__A , 0 , len(__A ) ) print(__A ) if __name__ == "__main__": main()
80
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) a__ : Union[str, Any] = {'configuration_mbart': ['MBART_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MBartConfig', 'MBartOnnxConfig']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : int = ['MBartTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : List[Any] = ['MBartTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : List[str] = [ 'MBART_PRETRAINED_MODEL_ARCHIVE_LIST', 'MBartForCausalLM', 'MBartForConditionalGeneration', 'MBartForQuestionAnswering', 'MBartForSequenceClassification', 'MBartModel', 'MBartPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : List[str] = [ 'TFMBartForConditionalGeneration', 'TFMBartModel', 'TFMBartPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : str = [ 'FlaxMBartForConditionalGeneration', 'FlaxMBartForQuestionAnswering', 'FlaxMBartForSequenceClassification', 'FlaxMBartModel', 'FlaxMBartPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mbart import MBART_PRETRAINED_CONFIG_ARCHIVE_MAP, MBartConfig, MBartOnnxConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mbart import MBartTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mbart_fast import MBartTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mbart import ( MBART_PRETRAINED_MODEL_ARCHIVE_LIST, MBartForCausalLM, MBartForConditionalGeneration, MBartForQuestionAnswering, MBartForSequenceClassification, MBartModel, MBartPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mbart import TFMBartForConditionalGeneration, TFMBartModel, TFMBartPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mbart import ( FlaxMBartForConditionalGeneration, FlaxMBartForQuestionAnswering, FlaxMBartForSequenceClassification, FlaxMBartModel, FlaxMBartPreTrainedModel, ) else: import sys a__ : Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
80
1
from ..utils import DummyObject, requires_backends class _UpperCamelCase ( metaclass=__lowercase ): '''simple docstring''' __UpperCamelCase : Union[str, Any] = ["torch", "scipy"] def __init__( self : int , *snake_case_ : int , **snake_case_ : int ): requires_backends(self , ["""torch""", """scipy"""] ) @classmethod def lowerCAmelCase__ ( cls : Any , *snake_case_ : List[Any] , **snake_case_ : str ): requires_backends(cls , ["""torch""", """scipy"""] ) @classmethod def lowerCAmelCase__ ( cls : Any , *snake_case_ : Union[str, Any] , **snake_case_ : Any ): requires_backends(cls , ["""torch""", """scipy"""] )
365
import json import multiprocessing import os import re from collections import defaultdict import torch from accelerate import Accelerator from accelerate.utils import set_seed from arguments import HumanEvalArguments from datasets import load_dataset, load_metric from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from tqdm import tqdm import transformers from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, StoppingCriteria, StoppingCriteriaList lowerCamelCase_ : str = ["""\nclass""", """\ndef""", """\n#""", """\n@""", """\nprint""", """\nif"""] class _UpperCamelCase ( _A ): '''simple docstring''' def __init__( self : Optional[int] , snake_case_ : List[str] , snake_case_ : Union[str, Any] , snake_case_ : Optional[Any]=None , snake_case_ : str=1 ): UpperCamelCase_: List[str] = tokenizer UpperCamelCase_: str = dataset UpperCamelCase_: List[str] = len(snake_case_ ) if n_tasks is None else n_tasks UpperCamelCase_: str = n_copies def __iter__( self : Tuple ): UpperCamelCase_: Optional[int] = [] for task in range(self.n_tasks ): # without strip, the model generate commented codes ... prompts.append(self.tokenizer.eos_token + self.dataset[task]["""prompt"""].strip() ) UpperCamelCase_: List[str] = self.tokenizer(snake_case_ , padding=snake_case_ , return_tensors="""pt""" ) for task in range(self.n_tasks ): for _ in range(self.n_copies ): yield { "ids": outputs.input_ids[task], "task_id": task, "input_len": outputs.attention_mask[task].sum(), } class _UpperCamelCase ( _A ): '''simple docstring''' def __init__( self : Any , snake_case_ : List[str] , snake_case_ : Dict , snake_case_ : int ): UpperCamelCase_: Union[str, Any] = start_length UpperCamelCase_: Dict = eof_strings UpperCamelCase_: List[str] = tokenizer def __call__( self : Tuple , snake_case_ : List[str] , snake_case_ : Optional[Any] , **snake_case_ : int ): UpperCamelCase_: Union[str, Any] = self.tokenizer.batch_decode(input_ids[:, self.start_length :] ) UpperCamelCase_: Dict = [] for decoded_generation in decoded_generations: done.append(any(stop_string in decoded_generation for stop_string in self.eof_strings ) ) return all(snake_case_ ) def A__ ( lowerCamelCase ) -> Optional[int]: UpperCamelCase_: str = re.split("""(%s)""" % """|""".join(lowerCamelCase ) , lowerCamelCase ) # last string should be "" return "".join(string_list[:-2] ) def A__ ( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase=20 , **lowerCamelCase ) -> int: UpperCamelCase_: str = defaultdict(lowerCamelCase ) # dict of list of generated tokens for step, batch in tqdm(enumerate(lowerCamelCase ) ): with torch.no_grad(): UpperCamelCase_: Optional[int] = batch["""ids"""].shape[-1] UpperCamelCase_: Dict = accelerator.unwrap_model(lowerCamelCase ).generate( input_ids=batch["""ids"""][:, : batch["""input_len"""]] , num_return_sequences=lowerCamelCase , **lowerCamelCase ) # each task is generated batch_size times UpperCamelCase_: Optional[int] = batch["""task_id"""].repeat(lowerCamelCase ) UpperCamelCase_: int = accelerator.pad_across_processes( lowerCamelCase , dim=1 , pad_index=tokenizer.pad_token_id ) UpperCamelCase_, UpperCamelCase_: Tuple = accelerator.gather((generated_tokens, generated_tasks) ) UpperCamelCase_: Tuple = generated_tokens.cpu().numpy() UpperCamelCase_: Any = generated_tasks.cpu().numpy() for task, generated_tokens in zip(lowerCamelCase , lowerCamelCase ): gen_token_dict[task].append(lowerCamelCase ) UpperCamelCase_: Dict = [[] for _ in range(lowerCamelCase )] for task, generated_tokens in gen_token_dict.items(): for s in generated_tokens: UpperCamelCase_: Any = tokenizer.decode(lowerCamelCase , skip_special_tokens=lowerCamelCase , clean_up_tokenization_spaces=lowerCamelCase ) code_gens[task].append(remove_last_block(lowerCamelCase ) ) return code_gens def A__ ( ) -> Union[str, Any]: # Setup configuration UpperCamelCase_: Optional[Any] = HfArgumentParser(lowerCamelCase ) UpperCamelCase_: str = parser.parse_args() transformers.logging.set_verbosity_error() # enables code execution in code_eval metric UpperCamelCase_: List[str] = args.HF_ALLOW_CODE_EVAL # make sure tokenizer plays nice with multiprocessing UpperCamelCase_: Union[str, Any] = """false""" if args.num_workers is None: UpperCamelCase_: Dict = multiprocessing.cpu_count() # Use dataset load to feed to accelerate UpperCamelCase_: List[Any] = Accelerator() set_seed(args.seed , device_specific=lowerCamelCase ) # Load model and tokenizer UpperCamelCase_: Any = AutoTokenizer.from_pretrained(args.model_ckpt ) UpperCamelCase_: Union[str, Any] = tokenizer.eos_token UpperCamelCase_: Optional[Any] = AutoModelForCausalLM.from_pretrained(args.model_ckpt ) # Generation settings UpperCamelCase_: Union[str, Any] = { """do_sample""": args.do_sample, """temperature""": args.temperature, """max_new_tokens""": args.max_new_tokens, """top_p""": args.top_p, """top_k""": args.top_k, """stopping_criteria""": StoppingCriteriaList([EndOfFunctionCriteria(0 , lowerCamelCase , lowerCamelCase )] ), } # Load evaluation dataset and metric UpperCamelCase_: Any = load_dataset("""openai_humaneval""" ) UpperCamelCase_: Union[str, Any] = load_metric("""code_eval""" ) UpperCamelCase_: Union[str, Any] = args.num_tasks if args.num_tasks is not None else len(human_eval["""test"""] ) UpperCamelCase_: List[Any] = args.n_samples // args.batch_size UpperCamelCase_: Any = TokenizedDataset(lowerCamelCase , human_eval["""test"""] , n_copies=lowerCamelCase , n_tasks=lowerCamelCase ) # do not confuse args.batch_size, which is actually the num_return_sequences UpperCamelCase_: Optional[int] = DataLoader(lowerCamelCase , batch_size=1 ) # Run a quick test to see if code evaluation is enabled try: UpperCamelCase_: List[str] = code_eval_metric.compute(references=[""""""] , predictions=[[""""""]] ) except ValueError as exception: print( """Code evaluation not enabled. Read the warning below carefully and then use `--HF_ALLOW_CODE_EVAL=\"1\"`""" """ flag to enable code evaluation.""" ) raise exception UpperCamelCase_, UpperCamelCase_: Dict = accelerator.prepare(lowerCamelCase , lowerCamelCase ) UpperCamelCase_: Union[str, Any] = complete_code( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , n_tasks=lowerCamelCase , batch_size=args.batch_size , **lowerCamelCase , ) if accelerator.is_main_process: UpperCamelCase_: List[Any] = [] for task in tqdm(range(lowerCamelCase ) ): UpperCamelCase_: Optional[Any] = human_eval["""test"""][task]["""test"""] UpperCamelCase_: Optional[int] = F'''check({human_eval["test"][task]["entry_point"]})''' references.append("""\n""" + test_func + """\n""" + entry_point ) # Evaluate completions with "code_eval" metric UpperCamelCase_, UpperCamelCase_: str = code_eval_metric.compute( references=lowerCamelCase , predictions=lowerCamelCase , num_workers=args.num_workers ) print(F'''Results: {pass_at_k}''' ) # Save results to json file with open(args.output_file , """w""" ) as fp: json.dump(lowerCamelCase , lowerCamelCase ) # For some reason the folliwng seems to be necessary sometimes for code_eval to work nice with multiprocessing # https://stackoverflow.com/questions/60804599/python-multiprocessing-keeps-spawning-the-whole-script if __name__ == "__main__": main()
223
0
import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def _UpperCAmelCase ( *SCREAMING_SNAKE_CASE__ : Union[str, Any] ): if not isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): __UpperCamelCase =list(SCREAMING_SNAKE_CASE__ ) for i in range(len(SCREAMING_SNAKE_CASE__ ) ): __UpperCamelCase =None gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() return objects def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : Exception ): __UpperCamelCase =[ 'CUDA out of memory.', # CUDA OOM 'cuDNN error: CUDNN_STATUS_NOT_SUPPORTED.', # CUDNN SNAFU 'DefaultCPUAllocator: can\'t allocate memory', # CPU OOM ] if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) and len(exception.args ) == 1: return any(err in exception.args[0] for err in _statements ) return False def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : callable = None , SCREAMING_SNAKE_CASE__ : int = 1_28 ): if function is None: return functools.partial(SCREAMING_SNAKE_CASE__ , starting_batch_size=SCREAMING_SNAKE_CASE__ ) __UpperCamelCase =starting_batch_size def decorator(*SCREAMING_SNAKE_CASE__ : Union[str, Any] , **SCREAMING_SNAKE_CASE__ : Optional[Any] ): nonlocal batch_size gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() __UpperCamelCase =list(inspect.signature(SCREAMING_SNAKE_CASE__ ).parameters.keys() ) # Guard against user error if len(SCREAMING_SNAKE_CASE__ ) < (len(SCREAMING_SNAKE_CASE__ ) + 1): __UpperCamelCase =', '.join([F'{arg}={value}' for arg, value in zip(params[1:] , args[1:] )] ) raise TypeError( F'Batch size was passed into `{function.__name__}` as the first argument when called.' F'Remove this as the decorator already does so: `{function.__name__}({arg_str})`' ) while True: if batch_size == 0: raise RuntimeError('No executable batch size found, reached zero.' ) try: return function(SCREAMING_SNAKE_CASE__ , *SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) except Exception as e: if should_reduce_batch_size(SCREAMING_SNAKE_CASE__ ): gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() batch_size //= 2 else: raise return decorator
62
from __future__ import annotations import math class UpperCAmelCase__ : """simple docstring""" def __init__( self , A_ ) -> None: __UpperCamelCase =size # approximate the overall size of segment tree with given value __UpperCamelCase =[0 for i in range(0 , 4 * size )] # create array to store lazy update __UpperCamelCase =[0 for i in range(0 , 4 * size )] __UpperCamelCase =[0 for i in range(0 , 4 * size )] # flag for lazy update def _a ( self , A_ ) -> int: return idx * 2 def _a ( self , A_ ) -> int: return idx * 2 + 1 def _a ( self , A_ , A_ , A_ , A_ ) -> None: if left_element == right_element: __UpperCamelCase =a[left_element - 1] else: __UpperCamelCase =(left_element + right_element) // 2 self.build(self.left(A_ ) , A_ , A_ , A_ ) self.build(self.right(A_ ) , mid + 1 , A_ , A_ ) __UpperCamelCase =max( self.segment_tree[self.left(A_ )] , self.segment_tree[self.right(A_ )] ) def _a ( self , A_ , A_ , A_ , A_ , A_ , A_ ) -> bool: if self.flag[idx] is True: __UpperCamelCase =self.lazy[idx] __UpperCamelCase =False if left_element != right_element: __UpperCamelCase =self.lazy[idx] __UpperCamelCase =self.lazy[idx] __UpperCamelCase =True __UpperCamelCase =True if right_element < a or left_element > b: return True if left_element >= a and right_element <= b: __UpperCamelCase =val if left_element != right_element: __UpperCamelCase =val __UpperCamelCase =val __UpperCamelCase =True __UpperCamelCase =True return True __UpperCamelCase =(left_element + right_element) // 2 self.update(self.left(A_ ) , A_ , A_ , A_ , A_ , A_ ) self.update(self.right(A_ ) , mid + 1 , A_ , A_ , A_ , A_ ) __UpperCamelCase =max( self.segment_tree[self.left(A_ )] , self.segment_tree[self.right(A_ )] ) return True def _a ( self , A_ , A_ , A_ , A_ , A_ ) -> int | float: if self.flag[idx] is True: __UpperCamelCase =self.lazy[idx] __UpperCamelCase =False if left_element != right_element: __UpperCamelCase =self.lazy[idx] __UpperCamelCase =self.lazy[idx] __UpperCamelCase =True __UpperCamelCase =True if right_element < a or left_element > b: return -math.inf if left_element >= a and right_element <= b: return self.segment_tree[idx] __UpperCamelCase =(left_element + right_element) // 2 __UpperCamelCase =self.query(self.left(A_ ) , A_ , A_ , A_ , A_ ) __UpperCamelCase =self.query(self.right(A_ ) , mid + 1 , A_ , A_ , A_ ) return max(A_ , A_ ) def __str__( self ) -> str: return str([self.query(1 , 1 , self.size , A_ , A_ ) for i in range(1 , self.size + 1 )] ) if __name__ == "__main__": _A = [1, 2, -4, 7, 3, -5, 6, 11, -20, 9, 14, 15, 5, 2, -8] _A = 15 _A = SegmentTree(size) segt.build(1, 1, size, A) print(segt.query(1, 1, size, 4, 6)) print(segt.query(1, 1, size, 7, 11)) print(segt.query(1, 1, size, 7, 12)) segt.update(1, 1, size, 1, 3, 111) print(segt.query(1, 1, size, 1, 15)) segt.update(1, 1, size, 7, 8, 235) print(segt)
62
1
"""simple docstring""" import numpy as np from sklearn.datasets import fetch_california_housing from sklearn.metrics import mean_absolute_error, mean_squared_error from sklearn.model_selection import train_test_split from xgboost import XGBRegressor def lowercase ( _snake_case : dict ) ->tuple: """simple docstring""" return (data["data"], data["target"]) def lowercase ( _snake_case : np.ndarray , _snake_case : np.ndarray , _snake_case : np.ndarray ) ->np.ndarray: """simple docstring""" __snake_case : Union[str, Any] = XGBRegressor(verbosity=0 , random_state=42 ) xgb.fit(_snake_case , _snake_case ) # Predict target for test data __snake_case : Any = xgb.predict(_snake_case ) __snake_case : Optional[int] = predictions.reshape(len(_snake_case ) , 1 ) return predictions def lowercase ( ) ->None: """simple docstring""" __snake_case : Union[str, Any] = fetch_california_housing() __snake_case , __snake_case : str = data_handling(_snake_case ) __snake_case , __snake_case , __snake_case , __snake_case : Dict = train_test_split( _snake_case , _snake_case , test_size=0.25 , random_state=1 ) __snake_case : List[str] = xgboost(_snake_case , _snake_case , _snake_case ) # Error printing print(f"""Mean Absolute Error : {mean_absolute_error(_snake_case , _snake_case )}""" ) print(f"""Mean Square Error : {mean_squared_error(_snake_case , _snake_case )}""" ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) main()
24
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE : Optional[Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE : Optional[int] = { """unc-nlp/lxmert-base-uncased""": """https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json""", } class _UpperCAmelCase ( __snake_case ): '''simple docstring''' lowerCamelCase__ ='lxmert' lowerCamelCase__ ={} def __init__(self , a_=3_05_22 , a_=7_68 , a_=12 , a_=95_00 , a_=16_00 , a_=4_00 , a_=30_72 , a_="gelu" , a_=0.1 , a_=0.1 , a_=5_12 , a_=2 , a_=0.02 , a_=1E-12 , a_=9 , a_=5 , a_=5 , a_=20_48 , a_=4 , a_=6.67 , a_=True , a_=True , a_=True , a_=True , a_=True , a_=True , a_=True , **a_ , ): '''simple docstring''' __snake_case : Optional[int] = vocab_size __snake_case : List[str] = hidden_size __snake_case : List[Any] = num_attention_heads __snake_case : int = hidden_act __snake_case : int = intermediate_size __snake_case : Any = hidden_dropout_prob __snake_case : List[Any] = attention_probs_dropout_prob __snake_case : Tuple = max_position_embeddings __snake_case : List[str] = type_vocab_size __snake_case : str = initializer_range __snake_case : Tuple = layer_norm_eps __snake_case : List[Any] = num_qa_labels __snake_case : int = num_object_labels __snake_case : Optional[Any] = num_attr_labels __snake_case : Union[str, Any] = l_layers __snake_case : Optional[int] = x_layers __snake_case : Optional[int] = r_layers __snake_case : Tuple = visual_feat_dim __snake_case : Optional[int] = visual_pos_dim __snake_case : Dict = visual_loss_normalizer __snake_case : str = task_matched __snake_case : Optional[Any] = task_mask_lm __snake_case : List[str] = task_obj_predict __snake_case : Optional[Any] = task_qa __snake_case : Any = visual_obj_loss __snake_case : int = visual_attr_loss __snake_case : List[Any] = visual_feat_loss __snake_case : Optional[Any] = {'''vision''': r_layers, '''cross_encoder''': x_layers, '''language''': l_layers} super().__init__(**a_ )
24
1
import gc import importlib.metadata import tempfile import unittest from packaging import version from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoTokenizer, BitsAndBytesConfig, pipeline, ) from transformers.testing_utils import ( is_torch_available, require_accelerate, require_bitsandbytes, require_torch, require_torch_gpu, require_torch_multi_gpu, slow, ) def A (__A : Union[str, Any] ) -> Tuple: """simple docstring""" if model.config.model_type == "gpt2": return model.transformer.h[0].mlp.c_fc return model.transformer.h[0].mlp.dense_ah_to_h if is_torch_available(): import torch import torch.nn as nn class __snake_case ( nn.Module ): def __init__( self : Union[str, Any] , _snake_case : nn.Module , _snake_case : int): """simple docstring""" super().__init__() UpperCAmelCase_ = module UpperCAmelCase_ = nn.Sequential( nn.Linear(module.in_features , UpperCamelCase_ , bias=UpperCamelCase_) , nn.Linear(UpperCamelCase_ , module.out_features , bias=UpperCamelCase_) , ) UpperCAmelCase_ = (2.0 / (5 * min(module.in_features , module.out_features))) ** 0.5 nn.init.normal_(self.adapter[0].weight , std=UpperCamelCase_) nn.init.zeros_(self.adapter[1].weight) self.adapter.to(module.weight.device) def lowerCamelCase ( self : Optional[int] , _snake_case : str , *_snake_case : int , **_snake_case : List[str]): """simple docstring""" return self.module(UpperCamelCase_ , *UpperCamelCase_ , **UpperCamelCase_) + self.adapter(UpperCamelCase_) @require_bitsandbytes @require_accelerate @require_torch @require_torch_gpu @slow class __snake_case ( unittest.TestCase ): # We keep the constants inside the init function and model loading inside setUp function # We need to test on relatively large models (aka >1b parameters otherwise the quantiztion may not work as expected) # Therefore here we use only bloom-1b3 to test our module UpperCAmelCase__ : Union[str, Any] = '''bigscience/bloom-1b7''' # Constant values UpperCAmelCase__ : List[Any] = 2.109_6595_5269_2574 UpperCAmelCase__ : List[str] = '''Hello my name is''' UpperCAmelCase__ : Union[str, Any] = set() EXPECTED_OUTPUTS.add('''Hello my name is John and I am a professional photographer. I''' ) EXPECTED_OUTPUTS.add('''Hello my name is John.\nI am a friend of your father.\n''' ) EXPECTED_OUTPUTS.add('''Hello my name is John Doe, I am a student at the University''' ) UpperCAmelCase__ : int = 1_0 def lowerCamelCase ( self : Any): """simple docstring""" UpperCAmelCase_ = AutoTokenizer.from_pretrained(self.model_name) class __snake_case ( UpperCamelCase__ ): def lowerCamelCase ( self : Tuple): """simple docstring""" super().setUp() # Models and tokenizer UpperCAmelCase_ = AutoModelForCausalLM.from_pretrained( self.model_name , torch_dtype=torch.floataa , device_map='''auto''') UpperCAmelCase_ = AutoModelForCausalLM.from_pretrained(self.model_name , load_in_abit=UpperCamelCase_ , device_map='''auto''') def lowerCamelCase ( self : Optional[Any]): """simple docstring""" del self.model_fpaa del self.model_abit gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self : List[str]): """simple docstring""" UpperCAmelCase_ = self.model_abit.config self.assertTrue(hasattr(UpperCamelCase_ , '''quantization_config''')) UpperCAmelCase_ = config.to_dict() UpperCAmelCase_ = config.to_diff_dict() UpperCAmelCase_ = config.to_json_string() def lowerCamelCase ( self : List[Any]): """simple docstring""" from bitsandbytes.nn import Paramsabit UpperCAmelCase_ = self.model_fpaa.get_memory_footprint() UpperCAmelCase_ = self.model_abit.get_memory_footprint() self.assertAlmostEqual(mem_fpaa / mem_abit , self.EXPECTED_RELATIVE_DIFFERENCE) UpperCAmelCase_ = get_some_linear_layer(self.model_abit) self.assertTrue(linear.weight.__class__ == Paramsabit) def lowerCamelCase ( self : Optional[Any]): """simple docstring""" from transformers import TaPreTrainedModel self.model_fpaa.get_memory_footprint() self.model_abit.get_memory_footprint() for name, module in self.model_abit.named_modules(): if isinstance(UpperCamelCase_ , torch.nn.Linear): if name not in ["lm_head"] + TaPreTrainedModel._keep_in_fpaa_modules: # 4-bit parameters are packed in uint8 variables self.assertTrue(module.weight.dtype == torch.uinta) def lowerCamelCase ( self : Dict): """simple docstring""" UpperCAmelCase_ = self.tokenizer(self.input_text , return_tensors='''pt''') UpperCAmelCase_ = self.model_abit.generate(input_ids=encoded_input['''input_ids'''].to(0) , max_new_tokens=10) self.assertIn(self.tokenizer.decode(output_sequences[0] , skip_special_tokens=UpperCamelCase_) , self.EXPECTED_OUTPUTS) def lowerCamelCase ( self : List[Any]): """simple docstring""" UpperCAmelCase_ = BitsAndBytesConfig() UpperCAmelCase_ = True UpperCAmelCase_ = AutoModelForCausalLM.from_pretrained( self.model_name , quantization_config=UpperCamelCase_ , device_map='''auto''') UpperCAmelCase_ = self.tokenizer(self.input_text , return_tensors='''pt''') UpperCAmelCase_ = model_abit_from_config.generate( input_ids=encoded_input['''input_ids'''].to(0) , max_new_tokens=10) self.assertIn(self.tokenizer.decode(output_sequences[0] , skip_special_tokens=UpperCamelCase_) , self.EXPECTED_OUTPUTS) def lowerCamelCase ( self : Union[str, Any]): """simple docstring""" with self.assertRaises(UpperCamelCase_), tempfile.TemporaryDirectory() as tmpdirname: self.model_abit.save_pretrained(UpperCamelCase_) def lowerCamelCase ( self : int): """simple docstring""" UpperCAmelCase_ = BitsAndBytesConfig() with self.assertRaises(UpperCamelCase_): UpperCAmelCase_ = AutoModelForCausalLM.from_pretrained( self.model_name , quantization_config=UpperCamelCase_ , load_in_abit=UpperCamelCase_ , device_map='''auto''' , bnb_abit_quant_type='''nf4''' , ) def lowerCamelCase ( self : List[str]): """simple docstring""" with self.assertRaises(UpperCamelCase_): # Tries with `str` self.model_abit.to('''cpu''') with self.assertRaises(UpperCamelCase_): # Tries with a `dtype`` self.model_abit.to(torch.floataa) with self.assertRaises(UpperCamelCase_): # Tries with a `device` self.model_abit.to(torch.device('''cuda:0''')) with self.assertRaises(UpperCamelCase_): # Tries with a `device` self.model_abit.float() with self.assertRaises(UpperCamelCase_): # Tries with a `device` self.model_abit.half() # Test if we did not break anything UpperCAmelCase_ = self.tokenizer(self.input_text , return_tensors='''pt''') UpperCAmelCase_ = self.model_fpaa.to(torch.floataa) UpperCAmelCase_ = self.model_fpaa.generate(input_ids=encoded_input['''input_ids'''].to(0) , max_new_tokens=10) # Check this does not throw an error UpperCAmelCase_ = self.model_fpaa.to('''cpu''') # Check this does not throw an error UpperCAmelCase_ = self.model_fpaa.half() # Check this does not throw an error UpperCAmelCase_ = self.model_fpaa.float() def lowerCamelCase ( self : str): """simple docstring""" UpperCAmelCase_ = AutoModelForSeqaSeqLM.from_pretrained('''t5-small''' , load_in_abit=UpperCamelCase_ , device_map='''auto''') self.assertTrue(model.decoder.block[0].layer[2].DenseReluDense.wo.weight.dtype == torch.floataa) @require_bitsandbytes @require_accelerate @require_torch @require_torch_gpu @slow class __snake_case ( unittest.TestCase ): @classmethod def lowerCamelCase ( cls : Any): """simple docstring""" UpperCAmelCase_ = '''t5-small''' UpperCAmelCase_ = '''google/flan-t5-small''' # flan-t5 uses dense-act instead of dense-relu-dense UpperCAmelCase_ = AutoTokenizer.from_pretrained(cls.model_name) UpperCAmelCase_ = '''Translate in German: Hello, my dog is cute''' def lowerCamelCase ( self : Optional[int]): """simple docstring""" gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self : Any): """simple docstring""" from transformers import TaForConditionalGeneration UpperCAmelCase_ = TaForConditionalGeneration._keep_in_fpaa_modules UpperCAmelCase_ = None # test with `t5-small` UpperCAmelCase_ = TaForConditionalGeneration.from_pretrained(self.model_name , load_in_abit=UpperCamelCase_ , device_map='''auto''') UpperCAmelCase_ = self.tokenizer(self.input_text , return_tensors='''pt''').to(0) UpperCAmelCase_ = model.generate(**UpperCamelCase_) # test with `flan-t5-small` UpperCAmelCase_ = TaForConditionalGeneration.from_pretrained( self.dense_act_model_name , load_in_abit=UpperCamelCase_ , device_map='''auto''') UpperCAmelCase_ = self.tokenizer(self.input_text , return_tensors='''pt''').to(0) UpperCAmelCase_ = model.generate(**UpperCamelCase_) UpperCAmelCase_ = modules def lowerCamelCase ( self : int): """simple docstring""" import bitsandbytes as bnb from transformers import TaForConditionalGeneration # test with `t5-small` UpperCAmelCase_ = TaForConditionalGeneration.from_pretrained(self.model_name , load_in_abit=UpperCamelCase_ , device_map='''auto''') # there was a bug with decoders - this test checks that it is fixed self.assertTrue(isinstance(model.decoder.block[0].layer[0].SelfAttention.q , bnb.nn.Linearabit)) UpperCAmelCase_ = self.tokenizer(self.input_text , return_tensors='''pt''').to(0) UpperCAmelCase_ = model.generate(**UpperCamelCase_) # test with `flan-t5-small` UpperCAmelCase_ = TaForConditionalGeneration.from_pretrained( self.dense_act_model_name , load_in_abit=UpperCamelCase_ , device_map='''auto''') UpperCAmelCase_ = self.tokenizer(self.input_text , return_tensors='''pt''').to(0) UpperCAmelCase_ = model.generate(**UpperCamelCase_) class __snake_case ( UpperCamelCase__ ): def lowerCamelCase ( self : Dict): """simple docstring""" super().setUp() # model_name UpperCAmelCase_ = '''bigscience/bloom-560m''' UpperCAmelCase_ = '''t5-small''' # Different types of model UpperCAmelCase_ = AutoModel.from_pretrained(self.model_name , load_in_abit=UpperCamelCase_ , device_map='''auto''') # Sequence classification model UpperCAmelCase_ = AutoModelForSequenceClassification.from_pretrained( self.model_name , load_in_abit=UpperCamelCase_ , device_map='''auto''') # CausalLM model UpperCAmelCase_ = AutoModelForCausalLM.from_pretrained(self.model_name , load_in_abit=UpperCamelCase_ , device_map='''auto''') # Seq2seq model UpperCAmelCase_ = AutoModelForSeqaSeqLM.from_pretrained( self.seq_to_seq_name , load_in_abit=UpperCamelCase_ , device_map='''auto''') def lowerCamelCase ( self : Any): """simple docstring""" del self.base_model del self.sequence_model del self.model_abit del self.seq_to_seq_model gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self : Union[str, Any]): """simple docstring""" from bitsandbytes.nn import Paramsabit self.assertTrue(self.base_model.h[-1].mlp.dense_ah_to_h.weight.__class__ == Paramsabit) # Other heads should be nn.Parameter self.assertTrue(self.model_abit.lm_head.weight.__class__ == torch.nn.Parameter) self.assertTrue(self.sequence_model.score.weight.__class__ == torch.nn.Parameter) self.assertTrue(self.seq_to_seq_model.lm_head.weight.__class__ == torch.nn.Parameter) class __snake_case ( UpperCamelCase__ ): def lowerCamelCase ( self : Dict): """simple docstring""" super().setUp() def lowerCamelCase ( self : Any): """simple docstring""" del self.pipe gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self : Optional[int]): """simple docstring""" UpperCAmelCase_ = pipeline( '''text-generation''' , model=self.model_name , model_kwargs={'''device_map''': '''auto''', '''load_in_4bit''': True, '''torch_dtype''': torch.floataa} , max_new_tokens=self.MAX_NEW_TOKENS , ) # Real second forward pass UpperCAmelCase_ = self.pipe(self.input_text) self.assertIn(pipeline_output[0]['''generated_text'''] , self.EXPECTED_OUTPUTS) @require_torch_multi_gpu class __snake_case ( UpperCamelCase__ ): def lowerCamelCase ( self : Dict): """simple docstring""" super().setUp() def lowerCamelCase ( self : Any): """simple docstring""" UpperCAmelCase_ = AutoModelForCausalLM.from_pretrained( self.model_name , load_in_abit=UpperCamelCase_ , device_map='''balanced''') # Check correct device map self.assertEqual(set(model_parallel.hf_device_map.values()) , {0, 1}) # Check that inference pass works on the model UpperCAmelCase_ = self.tokenizer(self.input_text , return_tensors='''pt''') # Second real batch UpperCAmelCase_ = model_parallel.generate(input_ids=encoded_input['''input_ids'''].to(0) , max_new_tokens=10) self.assertIn(self.tokenizer.decode(output_parallel[0] , skip_special_tokens=UpperCamelCase_) , self.EXPECTED_OUTPUTS) class __snake_case ( UpperCamelCase__ ): def lowerCamelCase ( self : str): """simple docstring""" UpperCAmelCase_ = '''facebook/opt-350m''' super().setUp() def lowerCamelCase ( self : Dict): """simple docstring""" if version.parse(importlib.metadata.version('''bitsandbytes''')) < version.parse('''0.37.0'''): return # Step 1: freeze all parameters UpperCAmelCase_ = AutoModelForCausalLM.from_pretrained(self.model_name , load_in_abit=UpperCamelCase_) self.assertEqual(set(model.hf_device_map.values()) , {torch.cuda.current_device()}) for param in model.parameters(): UpperCAmelCase_ = False # freeze the model - train adapters later if param.ndim == 1: # cast the small parameters (e.g. layernorm) to fp32 for stability UpperCAmelCase_ = param.data.to(torch.floataa) # Step 2: add adapters for _, module in model.named_modules(): if "OPTAttention" in repr(type(UpperCamelCase_)): UpperCAmelCase_ = LoRALayer(module.q_proj , rank=16) UpperCAmelCase_ = LoRALayer(module.k_proj , rank=16) UpperCAmelCase_ = LoRALayer(module.v_proj , rank=16) # Step 3: dummy batch UpperCAmelCase_ = self.tokenizer('''Test batch ''' , return_tensors='''pt''').to(0) # Step 4: Check if the gradient is not None with torch.cuda.amp.autocast(): UpperCAmelCase_ = model.forward(**UpperCamelCase_) out.logits.norm().backward() for module in model.modules(): if isinstance(UpperCamelCase_ , UpperCamelCase_): self.assertTrue(module.adapter[1].weight.grad is not None) self.assertTrue(module.adapter[1].weight.grad.norm().item() > 0) elif isinstance(UpperCamelCase_ , nn.Embedding): self.assertTrue(module.weight.grad is None) class __snake_case ( UpperCamelCase__ ): UpperCAmelCase__ : str = '''gpt2-xl''' UpperCAmelCase__ : Optional[int] = 3.3191_8548_5415_2187
51
from __future__ import annotations from collections.abc import Iterator class _a : def __init__( self: List[str] , UpperCamelCase_: int ) -> None: """simple docstring""" lowercase__ = value lowercase__ = None lowercase__ = None class _a : def __init__( self: Union[str, Any] , UpperCamelCase_: Node ) -> None: """simple docstring""" lowercase__ = tree def lowerCamelCase_ ( self: Any , UpperCamelCase_: Node | None ) -> int: """simple docstring""" if node is None: return 0 return node.value + ( self.depth_first_search(node.left ) + self.depth_first_search(node.right ) ) def __iter__( self: List[str] ) -> Iterator[int]: """simple docstring""" yield self.depth_first_search(self.tree ) if __name__ == "__main__": import doctest doctest.testmod()
110
0
import argparse import json import os import sys import tempfile import unittest from argparse import Namespace from dataclasses import dataclass, field from enum import Enum from pathlib import Path from typing import List, Literal, Optional import yaml from transformers import HfArgumentParser, TrainingArguments from transformers.hf_argparser import make_choice_type_function, string_to_bool # Since Python 3.10, we can use the builtin `|` operator for Union types # See PEP 604: https://peps.python.org/pep-0604 lowercase : Optional[Any] = sys.version_info >= (3, 10) def UpperCAmelCase_ (_lowerCAmelCase : Dict=None , _lowerCAmelCase : Any=None ): return field(default_factory=lambda: default , metadata=_lowerCAmelCase ) @dataclass class SCREAMING_SNAKE_CASE__ : """simple docstring""" lowercase : int lowercase : float lowercase : str lowercase : bool @dataclass class SCREAMING_SNAKE_CASE__ : """simple docstring""" lowercase : int = 42 lowercase : str = field(default='toto' , metadata={'help': 'help message'} ) @dataclass class SCREAMING_SNAKE_CASE__ : """simple docstring""" lowercase : bool = False lowercase : bool = True lowercase : Optional[bool] = None class SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ): """simple docstring""" lowercase : str = 'titi' lowercase : str = 'toto' class SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ): """simple docstring""" lowercase : str = 'titi' lowercase : Any = 'toto' lowercase : Dict = 42 @dataclass class SCREAMING_SNAKE_CASE__ : """simple docstring""" lowercase : BasicEnum = "toto" def __lowerCamelCase ( self ) -> Optional[Any]: '''simple docstring''' __UpperCamelCase : Optional[Any] = BasicEnum(self.foo ) @dataclass class SCREAMING_SNAKE_CASE__ : """simple docstring""" lowercase : MixedTypeEnum = "toto" def __lowerCamelCase ( self ) -> str: '''simple docstring''' __UpperCamelCase : str = MixedTypeEnum(self.foo ) @dataclass class SCREAMING_SNAKE_CASE__ : """simple docstring""" lowercase : Optional[int] = None lowercase : Optional[float] = field(default=lowerCamelCase__ , metadata={'help': 'help message'} ) lowercase : Optional[str] = None lowercase : Optional[List[str]] = list_field(default=[] ) lowercase : Optional[List[int]] = list_field(default=[] ) @dataclass class SCREAMING_SNAKE_CASE__ : """simple docstring""" lowercase : List[int] = list_field(default=[] ) lowercase : List[int] = list_field(default=[1, 2, 3] ) lowercase : List[str] = list_field(default=['Hallo', 'Bonjour', 'Hello'] ) lowercase : List[float] = list_field(default=[0.1, 0.2, 0.3] ) @dataclass class SCREAMING_SNAKE_CASE__ : """simple docstring""" lowercase : List[int] = field() lowercase : str = field() lowercase : BasicEnum = field() def __lowerCamelCase ( self ) -> List[Any]: '''simple docstring''' __UpperCamelCase : Dict = BasicEnum(self.required_enum ) @dataclass class SCREAMING_SNAKE_CASE__ : """simple docstring""" lowercase : int lowercase : "BasicEnum" = field() lowercase : "Optional[bool]" = None lowercase : "str" = field(default='toto' , metadata={'help': 'help message'} ) lowercase : "List[str]" = list_field(default=['Hallo', 'Bonjour', 'Hello'] ) if is_python_no_less_than_3_10: @dataclass class SCREAMING_SNAKE_CASE__ : """simple docstring""" lowercase : bool = False lowercase : bool = True lowercase : bool | None = None @dataclass class SCREAMING_SNAKE_CASE__ : """simple docstring""" lowercase : int | None = None lowercase : float | None = field(default=lowerCamelCase__ , metadata={'help': 'help message'} ) lowercase : str | None = None lowercase : list[str] | None = list_field(default=[] ) lowercase : list[int] | None = list_field(default=[] ) class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): """simple docstring""" def __lowerCamelCase ( self , __UpperCamelCase , __UpperCamelCase ) -> Optional[int]: '''simple docstring''' self.assertEqual(len(a._actions ) , len(b._actions ) ) for x, y in zip(a._actions , b._actions ): __UpperCamelCase : Union[str, Any] = {k: v for k, v in vars(__UpperCamelCase ).items() if k != "container"} __UpperCamelCase : Any = {k: v for k, v in vars(__UpperCamelCase ).items() if k != "container"} # Choices with mixed type have custom function as "type" # So we need to compare results directly for equality if xx.get("choices" , __UpperCamelCase ) and yy.get("choices" , __UpperCamelCase ): for expected_choice in yy["choices"] + xx["choices"]: self.assertEqual(xx["type"](__UpperCamelCase ) , yy["type"](__UpperCamelCase ) ) del xx["type"], yy["type"] self.assertEqual(__UpperCamelCase , __UpperCamelCase ) def __lowerCamelCase ( self ) -> Any: '''simple docstring''' __UpperCamelCase : Dict = HfArgumentParser(__UpperCamelCase ) __UpperCamelCase : Union[str, Any] = argparse.ArgumentParser() expected.add_argument("--foo" , type=__UpperCamelCase , required=__UpperCamelCase ) expected.add_argument("--bar" , type=__UpperCamelCase , required=__UpperCamelCase ) expected.add_argument("--baz" , type=__UpperCamelCase , required=__UpperCamelCase ) expected.add_argument("--flag" , type=__UpperCamelCase , default=__UpperCamelCase , const=__UpperCamelCase , nargs="?" ) self.argparsersEqual(__UpperCamelCase , __UpperCamelCase ) __UpperCamelCase : str = ["--foo", "1", "--baz", "quux", "--bar", "0.5"] ((__UpperCamelCase) , ) : Any = parser.parse_args_into_dataclasses(__UpperCamelCase , look_for_args_file=__UpperCamelCase ) self.assertFalse(example.flag ) def __lowerCamelCase ( self ) -> List[str]: '''simple docstring''' __UpperCamelCase : List[str] = HfArgumentParser(__UpperCamelCase ) __UpperCamelCase : int = argparse.ArgumentParser() expected.add_argument("--foo" , default=42 , type=__UpperCamelCase ) expected.add_argument("--baz" , default="toto" , type=__UpperCamelCase , help="help message" ) self.argparsersEqual(__UpperCamelCase , __UpperCamelCase ) def __lowerCamelCase ( self ) -> int: '''simple docstring''' __UpperCamelCase : Union[str, Any] = argparse.ArgumentParser() expected.add_argument("--foo" , type=__UpperCamelCase , default=__UpperCamelCase , const=__UpperCamelCase , nargs="?" ) expected.add_argument("--baz" , type=__UpperCamelCase , default=__UpperCamelCase , const=__UpperCamelCase , nargs="?" ) # A boolean no_* argument always has to come after its "default: True" regular counter-part # and its default must be set to False expected.add_argument("--no_baz" , action="store_false" , default=__UpperCamelCase , dest="baz" ) expected.add_argument("--opt" , type=__UpperCamelCase , default=__UpperCamelCase ) __UpperCamelCase : str = [WithDefaultBoolExample] if is_python_no_less_than_3_10: dataclass_types.append(__UpperCamelCase ) for dataclass_type in dataclass_types: __UpperCamelCase : Union[str, Any] = HfArgumentParser(__UpperCamelCase ) self.argparsersEqual(__UpperCamelCase , __UpperCamelCase ) __UpperCamelCase : Dict = parser.parse_args([] ) self.assertEqual(__UpperCamelCase , Namespace(foo=__UpperCamelCase , baz=__UpperCamelCase , opt=__UpperCamelCase ) ) __UpperCamelCase : Optional[Any] = parser.parse_args(["--foo", "--no_baz"] ) self.assertEqual(__UpperCamelCase , Namespace(foo=__UpperCamelCase , baz=__UpperCamelCase , opt=__UpperCamelCase ) ) __UpperCamelCase : Optional[int] = parser.parse_args(["--foo", "--baz"] ) self.assertEqual(__UpperCamelCase , Namespace(foo=__UpperCamelCase , baz=__UpperCamelCase , opt=__UpperCamelCase ) ) __UpperCamelCase : Optional[int] = parser.parse_args(["--foo", "True", "--baz", "True", "--opt", "True"] ) self.assertEqual(__UpperCamelCase , Namespace(foo=__UpperCamelCase , baz=__UpperCamelCase , opt=__UpperCamelCase ) ) __UpperCamelCase : Tuple = parser.parse_args(["--foo", "False", "--baz", "False", "--opt", "False"] ) self.assertEqual(__UpperCamelCase , Namespace(foo=__UpperCamelCase , baz=__UpperCamelCase , opt=__UpperCamelCase ) ) def __lowerCamelCase ( self ) -> Union[str, Any]: '''simple docstring''' __UpperCamelCase : Optional[int] = HfArgumentParser(__UpperCamelCase ) __UpperCamelCase : Any = argparse.ArgumentParser() expected.add_argument( "--foo" , default="toto" , choices=["titi", "toto", 42] , type=make_choice_type_function(["titi", "toto", 42] ) , ) self.argparsersEqual(__UpperCamelCase , __UpperCamelCase ) __UpperCamelCase : Tuple = parser.parse_args([] ) self.assertEqual(args.foo , "toto" ) __UpperCamelCase : Any = parser.parse_args_into_dataclasses([] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.toto ) __UpperCamelCase : int = parser.parse_args(["--foo", "titi"] ) self.assertEqual(args.foo , "titi" ) __UpperCamelCase : Optional[Any] = parser.parse_args_into_dataclasses(["--foo", "titi"] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.titi ) __UpperCamelCase : Any = parser.parse_args(["--foo", "42"] ) self.assertEqual(args.foo , 42 ) __UpperCamelCase : Tuple = parser.parse_args_into_dataclasses(["--foo", "42"] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.fourtytwo ) def __lowerCamelCase ( self ) -> Optional[int]: '''simple docstring''' @dataclass class SCREAMING_SNAKE_CASE__ : """simple docstring""" lowercase : Literal["titi", "toto", 42] = "toto" __UpperCamelCase : Any = HfArgumentParser(__UpperCamelCase ) __UpperCamelCase : List[str] = argparse.ArgumentParser() expected.add_argument( "--foo" , default="toto" , choices=("titi", "toto", 42) , type=make_choice_type_function(["titi", "toto", 42] ) , ) self.argparsersEqual(__UpperCamelCase , __UpperCamelCase ) __UpperCamelCase : Dict = parser.parse_args([] ) self.assertEqual(args.foo , "toto" ) __UpperCamelCase : Any = parser.parse_args(["--foo", "titi"] ) self.assertEqual(args.foo , "titi" ) __UpperCamelCase : Optional[Any] = parser.parse_args(["--foo", "42"] ) self.assertEqual(args.foo , 42 ) def __lowerCamelCase ( self ) -> Union[str, Any]: '''simple docstring''' __UpperCamelCase : Optional[int] = HfArgumentParser(__UpperCamelCase ) __UpperCamelCase : List[str] = argparse.ArgumentParser() expected.add_argument("--foo_int" , nargs="+" , default=[] , type=__UpperCamelCase ) expected.add_argument("--bar_int" , nargs="+" , default=[1, 2, 3] , type=__UpperCamelCase ) expected.add_argument("--foo_str" , nargs="+" , default=["Hallo", "Bonjour", "Hello"] , type=__UpperCamelCase ) expected.add_argument("--foo_float" , nargs="+" , default=[0.1, 0.2, 0.3] , type=__UpperCamelCase ) self.argparsersEqual(__UpperCamelCase , __UpperCamelCase ) __UpperCamelCase : List[Any] = parser.parse_args([] ) self.assertEqual( __UpperCamelCase , Namespace(foo_int=[] , bar_int=[1, 2, 3] , foo_str=["Hallo", "Bonjour", "Hello"] , foo_float=[0.1, 0.2, 0.3] ) , ) __UpperCamelCase : Optional[Any] = parser.parse_args("--foo_int 1 --bar_int 2 3 --foo_str a b c --foo_float 0.1 0.7".split() ) self.assertEqual(__UpperCamelCase , Namespace(foo_int=[1] , bar_int=[2, 3] , foo_str=["a", "b", "c"] , foo_float=[0.1, 0.7] ) ) def __lowerCamelCase ( self ) -> Optional[Any]: '''simple docstring''' __UpperCamelCase : int = argparse.ArgumentParser() expected.add_argument("--foo" , default=__UpperCamelCase , type=__UpperCamelCase ) expected.add_argument("--bar" , default=__UpperCamelCase , type=__UpperCamelCase , help="help message" ) expected.add_argument("--baz" , default=__UpperCamelCase , type=__UpperCamelCase ) expected.add_argument("--ces" , nargs="+" , default=[] , type=__UpperCamelCase ) expected.add_argument("--des" , nargs="+" , default=[] , type=__UpperCamelCase ) __UpperCamelCase : str = [OptionalExample] if is_python_no_less_than_3_10: dataclass_types.append(__UpperCamelCase ) for dataclass_type in dataclass_types: __UpperCamelCase : Optional[int] = HfArgumentParser(__UpperCamelCase ) self.argparsersEqual(__UpperCamelCase , __UpperCamelCase ) __UpperCamelCase : Optional[Any] = parser.parse_args([] ) self.assertEqual(__UpperCamelCase , Namespace(foo=__UpperCamelCase , bar=__UpperCamelCase , baz=__UpperCamelCase , ces=[] , des=[] ) ) __UpperCamelCase : int = parser.parse_args("--foo 12 --bar 3.14 --baz 42 --ces a b c --des 1 2 3".split() ) self.assertEqual(__UpperCamelCase , Namespace(foo=12 , bar=3.14 , baz="42" , ces=["a", "b", "c"] , des=[1, 2, 3] ) ) def __lowerCamelCase ( self ) -> int: '''simple docstring''' __UpperCamelCase : Any = HfArgumentParser(__UpperCamelCase ) __UpperCamelCase : List[Any] = argparse.ArgumentParser() expected.add_argument("--required_list" , nargs="+" , type=__UpperCamelCase , required=__UpperCamelCase ) expected.add_argument("--required_str" , type=__UpperCamelCase , required=__UpperCamelCase ) expected.add_argument( "--required_enum" , type=make_choice_type_function(["titi", "toto"] ) , choices=["titi", "toto"] , required=__UpperCamelCase , ) self.argparsersEqual(__UpperCamelCase , __UpperCamelCase ) def __lowerCamelCase ( self ) -> List[str]: '''simple docstring''' __UpperCamelCase : Tuple = HfArgumentParser(__UpperCamelCase ) __UpperCamelCase : Union[str, Any] = argparse.ArgumentParser() expected.add_argument("--foo" , type=__UpperCamelCase , required=__UpperCamelCase ) expected.add_argument( "--required_enum" , type=make_choice_type_function(["titi", "toto"] ) , choices=["titi", "toto"] , required=__UpperCamelCase , ) expected.add_argument("--opt" , type=__UpperCamelCase , default=__UpperCamelCase ) expected.add_argument("--baz" , default="toto" , type=__UpperCamelCase , help="help message" ) expected.add_argument("--foo_str" , nargs="+" , default=["Hallo", "Bonjour", "Hello"] , type=__UpperCamelCase ) self.argparsersEqual(__UpperCamelCase , __UpperCamelCase ) def __lowerCamelCase ( self ) -> str: '''simple docstring''' __UpperCamelCase : Any = HfArgumentParser(__UpperCamelCase ) __UpperCamelCase : List[str] = { "foo": 12, "bar": 3.14, "baz": "42", "flag": True, } __UpperCamelCase : Optional[int] = parser.parse_dict(__UpperCamelCase )[0] __UpperCamelCase : Dict = BasicExample(**__UpperCamelCase ) self.assertEqual(__UpperCamelCase , __UpperCamelCase ) def __lowerCamelCase ( self ) -> Optional[Any]: '''simple docstring''' __UpperCamelCase : Optional[Any] = HfArgumentParser(__UpperCamelCase ) __UpperCamelCase : Any = { "foo": 12, "bar": 3.14, "baz": "42", "flag": True, "extra": 42, } self.assertRaises(__UpperCamelCase , parser.parse_dict , __UpperCamelCase , allow_extra_keys=__UpperCamelCase ) def __lowerCamelCase ( self ) -> Tuple: '''simple docstring''' __UpperCamelCase : Dict = HfArgumentParser(__UpperCamelCase ) __UpperCamelCase : int = { "foo": 12, "bar": 3.14, "baz": "42", "flag": True, } with tempfile.TemporaryDirectory() as tmp_dir: __UpperCamelCase : List[Any] = os.path.join(__UpperCamelCase , "temp_json" ) os.mkdir(__UpperCamelCase ) with open(temp_local_path + ".json" , "w+" ) as f: json.dump(__UpperCamelCase , __UpperCamelCase ) __UpperCamelCase : List[str] = parser.parse_yaml_file(Path(temp_local_path + ".json" ) )[0] __UpperCamelCase : int = BasicExample(**__UpperCamelCase ) self.assertEqual(__UpperCamelCase , __UpperCamelCase ) def __lowerCamelCase ( self ) -> Tuple: '''simple docstring''' __UpperCamelCase : List[Any] = HfArgumentParser(__UpperCamelCase ) __UpperCamelCase : Tuple = { "foo": 12, "bar": 3.14, "baz": "42", "flag": True, } with tempfile.TemporaryDirectory() as tmp_dir: __UpperCamelCase : Any = os.path.join(__UpperCamelCase , "temp_yaml" ) os.mkdir(__UpperCamelCase ) with open(temp_local_path + ".yaml" , "w+" ) as f: yaml.dump(__UpperCamelCase , __UpperCamelCase ) __UpperCamelCase : Optional[int] = parser.parse_yaml_file(Path(temp_local_path + ".yaml" ) )[0] __UpperCamelCase : Optional[Any] = BasicExample(**__UpperCamelCase ) self.assertEqual(__UpperCamelCase , __UpperCamelCase ) def __lowerCamelCase ( self ) -> int: '''simple docstring''' __UpperCamelCase : Optional[Any] = HfArgumentParser(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase )
171
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase : Union[str, Any] = { "configuration_mctct": ["MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP", "MCTCTConfig"], "feature_extraction_mctct": ["MCTCTFeatureExtractor"], "processing_mctct": ["MCTCTProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Tuple = [ "MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST", "MCTCTForCTC", "MCTCTModel", "MCTCTPreTrainedModel", ] if TYPE_CHECKING: from .configuration_mctct import MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP, MCTCTConfig from .feature_extraction_mctct import MCTCTFeatureExtractor from .processing_mctct import MCTCTProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mctct import MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST, MCTCTForCTC, MCTCTModel, MCTCTPreTrainedModel else: import sys lowercase : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
171
1
'''simple docstring''' import copy import inspect import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import TimesformerConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, TimesformerForVideoClassification, TimesformerModel, ) from transformers.models.timesformer.modeling_timesformer import TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from transformers import VideoMAEImageProcessor class a__ : """simple docstring""" def __init__(self , __lowercase , __lowercase=13 , __lowercase=10 , __lowercase=3 , __lowercase=2 , __lowercase=2 , __lowercase=True , __lowercase=True , __lowercase=32 , __lowercase=5 , __lowercase=4 , __lowercase=37 , __lowercase="gelu" , __lowercase=0.1 , __lowercase=0.1 , __lowercase=10 , __lowercase=0.0_2 , __lowercase="divided_space_time" , __lowercase=None , ): __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = image_size __lowerCAmelCase = num_channels __lowerCAmelCase = patch_size __lowerCAmelCase = num_frames __lowerCAmelCase = is_training __lowerCAmelCase = use_labels __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = attention_type __lowerCAmelCase = initializer_range __lowerCAmelCase = scope __lowerCAmelCase = num_labels # in TimeSformer, the number of spatial tokens equals num_frames * num_patches per frame + 1 CLS token __lowerCAmelCase = (image_size // patch_size) ** 2 __lowerCAmelCase = (num_frames) * self.num_patches_per_frame + 1 def _snake_case (self ): __lowerCAmelCase = floats_tensor( [self.batch_size, self.num_frames, self.num_channels, self.image_size, self.image_size] ) __lowerCAmelCase = None if self.use_labels: __lowerCAmelCase = ids_tensor([self.batch_size] , self.num_labels ) __lowerCAmelCase = self.get_config() return config, pixel_values, labels def _snake_case (self ): __lowerCAmelCase = TimesformerConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_frames=self.num_frames , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , attention_type=self.attention_type , ) __lowerCAmelCase = self.num_labels return config def _snake_case (self , __lowercase , __lowercase , __lowercase ): __lowerCAmelCase = TimesformerModel(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() __lowerCAmelCase = model(__lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _snake_case (self , __lowercase , __lowercase , __lowercase ): __lowerCAmelCase = TimesformerForVideoClassification(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() __lowerCAmelCase = model(__lowerCAmelCase ) # verify the logits shape __lowerCAmelCase = torch.Size((self.batch_size, self.num_labels) ) self.parent.assertEqual(result.logits.shape , __lowerCAmelCase ) def _snake_case (self ): __lowerCAmelCase = self.prepare_config_and_inputs() __lowerCAmelCase = config_and_inputs __lowerCAmelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class a__ ( a__ , a__ , unittest.TestCase ): """simple docstring""" __UpperCamelCase : Any = (TimesformerModel, TimesformerForVideoClassification) if is_torch_available() else () __UpperCamelCase : Tuple = ( {'feature-extraction': TimesformerModel, 'video-classification': TimesformerForVideoClassification} if is_torch_available() else {} ) __UpperCamelCase : Union[str, Any] = False __UpperCamelCase : str = False __UpperCamelCase : Optional[Any] = False __UpperCamelCase : Any = False def _snake_case (self ): __lowerCAmelCase = TimesformerModelTester(self ) __lowerCAmelCase = ConfigTester( self , config_class=__lowerCAmelCase , has_text_modality=__lowerCAmelCase , hidden_size=37 ) def _snake_case (self , __lowercase , __lowercase , __lowercase=False ): __lowerCAmelCase = copy.deepcopy(__lowerCAmelCase ) if return_labels: if model_class in get_values(__lowerCAmelCase ): __lowerCAmelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__lowerCAmelCase ) return inputs_dict def _snake_case (self ): self.config_tester.run_common_tests() @unittest.skip(reason='''TimeSformer does not use inputs_embeds''' ) def _snake_case (self ): pass def _snake_case (self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCAmelCase = model_class(__lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __lowerCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCAmelCase , nn.Linear ) ) def _snake_case (self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCAmelCase = model_class(__lowerCAmelCase ) __lowerCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCAmelCase = [*signature.parameters.keys()] __lowerCAmelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , __lowerCAmelCase ) def _snake_case (self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCAmelCase ) def _snake_case (self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_video_classification(*__lowerCAmelCase ) @slow def _snake_case (self ): for model_name in TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase = TimesformerModel.from_pretrained(__lowerCAmelCase ) self.assertIsNotNone(__lowerCAmelCase ) def _snake_case (self ): if not self.has_attentions: pass else: __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() __lowerCAmelCase = True for model_class in self.all_model_classes: __lowerCAmelCase = self.model_tester.seq_length __lowerCAmelCase = self.model_tester.num_frames __lowerCAmelCase = True __lowerCAmelCase = False __lowerCAmelCase = True __lowerCAmelCase = model_class(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() with torch.no_grad(): __lowerCAmelCase = model(**self._prepare_for_class(__lowerCAmelCase , __lowerCAmelCase ) ) __lowerCAmelCase = outputs.attentions self.assertEqual(len(__lowerCAmelCase ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] __lowerCAmelCase = True __lowerCAmelCase = model_class(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() with torch.no_grad(): __lowerCAmelCase = model(**self._prepare_for_class(__lowerCAmelCase , __lowerCAmelCase ) ) __lowerCAmelCase = outputs.attentions self.assertEqual(len(__lowerCAmelCase ) , self.model_tester.num_hidden_layers ) # attentions has shape (batch_size x num_frames) x num_heads x (num_patches per frame + 1) x (num_patches per frame + 1) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len // num_frames + 1, seq_len // num_frames + 1] , ) __lowerCAmelCase = len(__lowerCAmelCase ) # Check attention is always last and order is fine __lowerCAmelCase = True __lowerCAmelCase = True __lowerCAmelCase = model_class(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() with torch.no_grad(): __lowerCAmelCase = model(**self._prepare_for_class(__lowerCAmelCase , __lowerCAmelCase ) ) self.assertEqual(out_len + 1 , len(__lowerCAmelCase ) ) __lowerCAmelCase = outputs.attentions self.assertEqual(len(__lowerCAmelCase ) , self.model_tester.num_hidden_layers ) # attentions has shape (batch_size x num_frames) x num_heads x (num_patches per frame + 1) x (num_patches per frame + 1) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len // num_frames + 1, seq_len // num_frames + 1] , ) def _snake_case (self ): def check_hidden_states_output(__lowercase , __lowercase , __lowercase ): __lowerCAmelCase = model_class(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() with torch.no_grad(): __lowerCAmelCase = model(**self._prepare_for_class(__lowerCAmelCase , __lowerCAmelCase ) ) __lowerCAmelCase = outputs.hidden_states __lowerCAmelCase = self.model_tester.num_hidden_layers + 1 self.assertEqual(len(__lowerCAmelCase ) , __lowerCAmelCase ) __lowerCAmelCase = self.model_tester.seq_length self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCAmelCase = True check_hidden_states_output(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowerCAmelCase = True check_hidden_states_output(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) def __magic_name__( ): __lowerCAmelCase = hf_hub_download( repo_id='''hf-internal-testing/spaghetti-video''', filename='''eating_spaghetti.npy''', repo_type='''dataset''') __lowerCAmelCase = np.load(lowerCamelCase) return list(lowerCamelCase) @require_torch @require_vision class a__ ( unittest.TestCase ): """simple docstring""" @cached_property def _snake_case (self ): # logits were tested with a different mean and std, so we use the same here return ( VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5] , image_std=[0.5, 0.5, 0.5] ) if is_vision_available() else None ) @slow def _snake_case (self ): __lowerCAmelCase = TimesformerForVideoClassification.from_pretrained('''facebook/timesformer-base-finetuned-k400''' ).to( __lowerCAmelCase ) __lowerCAmelCase = self.default_image_processor __lowerCAmelCase = prepare_video() __lowerCAmelCase = image_processor(video[:8] , return_tensors='''pt''' ).to(__lowerCAmelCase ) # forward pass with torch.no_grad(): __lowerCAmelCase = model(**__lowerCAmelCase ) # verify the logits __lowerCAmelCase = torch.Size((1, 4_00) ) self.assertEqual(outputs.logits.shape , __lowerCAmelCase ) __lowerCAmelCase = torch.tensor([-0.3_0_1_6, -0.7_7_1_3, -0.4_2_0_5] ).to(__lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __lowerCAmelCase , atol=1e-4 ) )
174
'''simple docstring''' def __UpperCamelCase ( UpperCAmelCase ): lowercase__ , lowercase__ : Optional[int] = [], [] while len(UpperCAmelCase ) > 1: lowercase__ , lowercase__ : List[str] = min(UpperCAmelCase ), max(UpperCAmelCase ) start.append(UpperCAmelCase ) end.append(UpperCAmelCase ) collection.remove(UpperCAmelCase ) collection.remove(UpperCAmelCase ) end.reverse() return start + collection + end if __name__ == "__main__": __a: Optional[int] = input("""Enter numbers separated by a comma:\n""").strip() __a: Tuple = [int(item) for item in user_input.split(""",""")] print(*merge_sort(unsorted), sep=""",""")
198
0
def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE ): A_ : str = [] if len(SCREAMING_SNAKE_CASE ) == 1: return [nums.copy()] for _ in range(len(SCREAMING_SNAKE_CASE ) ): A_ : int = nums.pop(0 ) A_ : int = permute(SCREAMING_SNAKE_CASE ) for perm in permutations: perm.append(SCREAMING_SNAKE_CASE ) result.extend(SCREAMING_SNAKE_CASE ) nums.append(SCREAMING_SNAKE_CASE ) return result def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE ): def backtrack(SCREAMING_SNAKE_CASE ): if start == len(SCREAMING_SNAKE_CASE ) - 1: output.append(nums[:] ) else: for i in range(SCREAMING_SNAKE_CASE , len(SCREAMING_SNAKE_CASE ) ): A_ , A_ : List[Any] = nums[i], nums[start] backtrack(start + 1 ) A_ , A_ : Optional[Any] = nums[i], nums[start] # backtrack A_ : Optional[int] = [] backtrack(0 ) return output if __name__ == "__main__": import doctest # use res to print the data in permute2 function UpperCamelCase = permutea([1, 2, 3]) print(res) doctest.testmod()
65
import math import random def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = False ): if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value UpperCamelCase = 0.02 def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): A_ : str = float(2 * (random.randint(1 , 100 )) - 1 ) for _ in range(SCREAMING_SNAKE_CASE ): # Forward propagation A_ : Optional[Any] = sigmoid_function(INITIAL_VALUE * weight ) # How much did we miss? A_ : Any = (expected / 100) - layer_a # Error delta A_ : List[str] = layer_1_error * sigmoid_function(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Update weight weight += INITIAL_VALUE * layer_1_delta return layer_a * 100 if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase = int(input("""Expected value: """)) UpperCamelCase = int(input("""Number of propagations: """)) print(forward_propagation(expected, number_propagations))
65
1
import datasets from .evaluate import evaluate _lowerCamelCase : Any = """\ @inproceedings{Rajpurkar2016SQuAD10, title={SQuAD: 100, 000+ Questions for Machine Comprehension of Text}, author={Pranav Rajpurkar and Jian Zhang and Konstantin Lopyrev and Percy Liang}, booktitle={EMNLP}, year={2016} } """ _lowerCamelCase : int = """ This metric wrap the official scoring script for version 1 of the Stanford Question Answering Dataset (SQuAD). Stanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by crowdworkers on a set of Wikipedia articles, where the answer to every question is a segment of text, or span, from the corresponding reading passage, or the question might be unanswerable. """ _lowerCamelCase : Dict = """ Computes SQuAD scores (F1 and EM). Args: predictions: List of question-answers dictionaries with the following key-values: - 'id': id of the question-answer pair as given in the references (see below) - 'prediction_text': the text of the answer references: List of question-answers dictionaries with the following key-values: - 'id': id of the question-answer pair (see above), - 'answers': a Dict in the SQuAD dataset format { 'text': list of possible texts for the answer, as a list of strings 'answer_start': list of start positions for the answer, as a list of ints } Note that answer_start values are not taken into account to compute the metric. Returns: 'exact_match': Exact match (the normalized answer exactly match the gold answer) 'f1': The F-score of predicted tokens versus the gold answer Examples: >>> predictions = [{'prediction_text': '1976', 'id': '56e10a3be3433e1400422b22'}] >>> references = [{'answers': {'answer_start': [97], 'text': ['1976']}, 'id': '56e10a3be3433e1400422b22'}] >>> squad_metric = datasets.load_metric(\"squad\") >>> results = squad_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 100.0, 'f1': 100.0} """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCamelCase_ ( datasets.Metric ): '''simple docstring''' def SCREAMING_SNAKE_CASE ( self : List[Any]) ->int: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': {'''id''': datasets.Value('''string'''), '''prediction_text''': datasets.Value('''string''')}, '''references''': { '''id''': datasets.Value('''string'''), '''answers''': datasets.features.Sequence( { '''text''': datasets.Value('''string'''), '''answer_start''': datasets.Value('''int32'''), }), }, }) , codebase_urls=['''https://rajpurkar.github.io/SQuAD-explorer/'''] , reference_urls=['''https://rajpurkar.github.io/SQuAD-explorer/'''] , ) def SCREAMING_SNAKE_CASE ( self : List[str] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Union[str, Any]) ->Tuple: '''simple docstring''' A__ = {prediction['''id''']: prediction['''prediction_text'''] for prediction in predictions} A__ = [ { '''paragraphs''': [ { '''qas''': [ { '''answers''': [{'''text''': answer_text} for answer_text in ref['''answers''']['''text''']], '''id''': ref['''id'''], } for ref in references ] } ] } ] A__ = evaluate(dataset=UpperCAmelCase__ , predictions=UpperCAmelCase__) return score
14
_lowerCamelCase : Optional[int] = 65521 def SCREAMING_SNAKE_CASE ( lowercase_ ) -> int: """simple docstring""" A__ = 1 A__ = 0 for plain_chr in plain_text: A__ = (a + ord(lowercase_ )) % MOD_ADLER A__ = (b + a) % MOD_ADLER return (b << 16) | a
14
1
import json import os import shutil import tempfile import unittest import numpy as np from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES, BertTokenizer from transformers.testing_utils import require_tokenizers, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import VisionTextDualEncoderProcessor, ViTImageProcessor @require_tokenizers @require_vision class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __magic_name__ ( self : Tuple ): """simple docstring""" _A: int = tempfile.mkdtemp() # fmt: off _A: List[str] = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest'''] # fmt: on _A: List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) _A: Tuple = { '''do_resize''': True, '''size''': {'''height''': 1_8, '''width''': 1_8}, '''do_normalize''': True, '''image_mean''': [0.5, 0.5, 0.5], '''image_std''': [0.5, 0.5, 0.5], } _A: List[str] = os.path.join(self.tmpdirname , lowerCAmelCase_ ) with open(self.image_processor_file , '''w''' , encoding='''utf-8''' ) as fp: json.dump(lowerCAmelCase_ , lowerCAmelCase_ ) def __magic_name__ ( self : Dict , **lowerCAmelCase_ : Tuple ): """simple docstring""" return BertTokenizer.from_pretrained(self.tmpdirname , **lowerCAmelCase_ ) def __magic_name__ ( self : Dict , **lowerCAmelCase_ : List[Any] ): """simple docstring""" return ViTImageProcessor.from_pretrained(self.tmpdirname , **lowerCAmelCase_ ) def __magic_name__ ( self : Optional[int] ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def __magic_name__ ( self : Tuple ): """simple docstring""" _A: Tuple = [np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta )] _A: int = [Image.fromarray(np.moveaxis(lowerCAmelCase_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def __magic_name__ ( self : Dict ): """simple docstring""" _A: List[str] = self.get_tokenizer() _A: List[Any] = self.get_image_processor() _A: Optional[int] = VisionTextDualEncoderProcessor(tokenizer=lowerCAmelCase_ , image_processor=lowerCAmelCase_ ) processor.save_pretrained(self.tmpdirname ) _A: Optional[int] = VisionTextDualEncoderProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCAmelCase_ ) def __magic_name__ ( self : Dict ): """simple docstring""" _A: Any = VisionTextDualEncoderProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _A: Union[str, Any] = self.get_tokenizer(bos_token='''(BOS)''' , eos_token='''(EOS)''' ) _A: List[str] = self.get_image_processor(do_normalize=lowerCAmelCase_ , padding_value=1.0 ) _A: Any = VisionTextDualEncoderProcessor.from_pretrained( self.tmpdirname , bos_token='''(BOS)''' , eos_token='''(EOS)''' , do_normalize=lowerCAmelCase_ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCAmelCase_ ) def __magic_name__ ( self : int ): """simple docstring""" _A: int = self.get_image_processor() _A: List[str] = self.get_tokenizer() _A: Tuple = VisionTextDualEncoderProcessor(tokenizer=lowerCAmelCase_ , image_processor=lowerCAmelCase_ ) _A: str = self.prepare_image_inputs() _A: str = image_processor(lowerCAmelCase_ , return_tensors='''np''' ) _A: Optional[Any] = processor(images=lowerCAmelCase_ , return_tensors='''np''' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def __magic_name__ ( self : List[Any] ): """simple docstring""" _A: Tuple = self.get_image_processor() _A: List[Any] = self.get_tokenizer() _A: Union[str, Any] = VisionTextDualEncoderProcessor(tokenizer=lowerCAmelCase_ , image_processor=lowerCAmelCase_ ) _A: Union[str, Any] = '''lower newer''' _A: Dict = processor(text=lowerCAmelCase_ ) _A: Tuple = tokenizer(lowerCAmelCase_ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def __magic_name__ ( self : Optional[Any] ): """simple docstring""" _A: Optional[int] = self.get_image_processor() _A: Dict = self.get_tokenizer() _A: int = VisionTextDualEncoderProcessor(tokenizer=lowerCAmelCase_ , image_processor=lowerCAmelCase_ ) _A: str = '''lower newer''' _A: Tuple = self.prepare_image_inputs() _A: Any = processor(text=lowerCAmelCase_ , images=lowerCAmelCase_ ) self.assertListEqual(list(inputs.keys() ) , ['''input_ids''', '''token_type_ids''', '''attention_mask''', '''pixel_values'''] ) # test if it raises when no input is passed with self.assertRaises(lowerCAmelCase_ ): processor() def __magic_name__ ( self : Tuple ): """simple docstring""" _A: Dict = self.get_image_processor() _A: Optional[int] = self.get_tokenizer() _A: str = VisionTextDualEncoderProcessor(tokenizer=lowerCAmelCase_ , image_processor=lowerCAmelCase_ ) _A: List[Any] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _A: Optional[Any] = processor.batch_decode(lowerCAmelCase_ ) _A: str = tokenizer.batch_decode(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def __magic_name__ ( self : str ): """simple docstring""" _A: List[Any] = self.get_image_processor() _A: Optional[Any] = self.get_tokenizer() _A: str = VisionTextDualEncoderProcessor(tokenizer=lowerCAmelCase_ , image_processor=lowerCAmelCase_ ) _A: Optional[int] = '''lower newer''' _A: str = self.prepare_image_inputs() _A: List[Any] = processor(text=lowerCAmelCase_ , images=lowerCAmelCase_ ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
301
from __future__ import annotations UpperCAmelCase__ : List[str] = list[list[int]] # assigning initial values to the grid UpperCAmelCase__ : Matrix = [ [3, 0, 6, 5, 0, 8, 4, 0, 0], [5, 2, 0, 0, 0, 0, 0, 0, 0], [0, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] # a grid with no solution UpperCAmelCase__ : Matrix = [ [5, 0, 6, 5, 0, 8, 4, 0, 3], [5, 2, 0, 0, 0, 0, 0, 0, 2], [1, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] def lowerCamelCase__ ( a , a , a , a ) -> bool: for i in range(9 ): if grid[row][i] == n or grid[i][column] == n: return False for i in range(3 ): for j in range(3 ): if grid[(row - row % 3) + i][(column - column % 3) + j] == n: return False return True def lowerCamelCase__ ( a ) -> tuple[int, int] | None: for i in range(9 ): for j in range(9 ): if grid[i][j] == 0: return i, j return None def lowerCamelCase__ ( a ) -> Matrix | None: if location := find_empty_location(a ): _A , _A: Optional[Any] = location else: # If the location is ``None``, then the grid is solved. return grid for digit in range(1 , 10 ): if is_safe(a , a , a , a ): _A: str = digit if sudoku(a ) is not None: return grid _A: Tuple = 0 return None def lowerCamelCase__ ( a ) -> None: for row in grid: for cell in row: print(a , end=''' ''' ) print() if __name__ == "__main__": # make a copy of grid so that you can compare with the unmodified grid for example_grid in (initial_grid, no_solution): print('\nExample grid:\n' + '=' * 20) print_solution(example_grid) print('\nExample grid solution:') UpperCAmelCase__ : int = sudoku(example_grid) if solution is not None: print_solution(solution) else: print('Cannot find a solution.')
301
1
from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING lowerCamelCase__ = logging.get_logger(__name__) @add_end_docstrings(__magic_name__ ) class A__ ( __magic_name__ ): def __init__( self : int , *a : Dict , **a : Union[str, Any] ): '''simple docstring''' super().__init__(*a , **a ) requires_backends(self , 'vision' ) self.check_model_type( TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING if self.framework == 'tf' else MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING ) def _lowerCamelCase ( self : Dict , a : List[str]=None ): '''simple docstring''' lowerCAmelCase__ : Any = {} if top_k is not None: lowerCAmelCase__ : Tuple = top_k return {}, {}, postprocess_params def __call__( self : Any , a : Union[str, List[str], "Image.Image", List["Image.Image"]] , **a : List[Any] ): '''simple docstring''' return super().__call__(a , **a ) def _lowerCamelCase ( self : Any , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = load_image(a ) lowerCAmelCase__ : Optional[int] = self.image_processor(images=a , return_tensors=self.framework ) return model_inputs def _lowerCamelCase ( self : Optional[int] , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.model(**a ) return model_outputs def _lowerCamelCase ( self : Optional[Any] , a : List[Any] , a : List[Any]=5 ): '''simple docstring''' if top_k > self.model.config.num_labels: lowerCAmelCase__ : Optional[int] = self.model.config.num_labels if self.framework == "pt": lowerCAmelCase__ : List[Any] = model_outputs.logits.softmax(-1 )[0] lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = probs.topk(a ) elif self.framework == "tf": lowerCAmelCase__ : Any = stable_softmax(model_outputs.logits , axis=-1 )[0] lowerCAmelCase__ : Any = tf.math.top_k(a , k=a ) lowerCAmelCase__ , lowerCAmelCase__ : int = topk.values.numpy(), topk.indices.numpy() else: raise ValueError(f'''Unsupported framework: {self.framework}''' ) lowerCAmelCase__ : List[Any] = scores.tolist() lowerCAmelCase__ : List[str] = ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(a , a )]
212
import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> np.array: lowerCAmelCase__ : Dict = F'''{sampling_rate}''' lowerCAmelCase__ : Any = '1' lowerCAmelCase__ : Optional[Any] = 'f32le' lowerCAmelCase__ : Any = [ 'ffmpeg', '-i', 'pipe:0', '-ac', ac, '-ar', ar, '-f', format_for_conversion, '-hide_banner', '-loglevel', 'quiet', 'pipe:1', ] try: with subprocess.Popen(SCREAMING_SNAKE_CASE_ , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: lowerCAmelCase__ : List[Any] = ffmpeg_process.communicate(SCREAMING_SNAKE_CASE_ ) except FileNotFoundError as error: raise ValueError('ffmpeg was not found but is required to load audio files from filename' ) from error lowerCAmelCase__ : List[str] = output_stream[0] lowerCAmelCase__ : str = np.frombuffer(SCREAMING_SNAKE_CASE_ , np.floataa ) if audio.shape[0] == 0: raise ValueError('Malformed soundfile' ) return audio def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = "f32le" , ) -> Dict: lowerCAmelCase__ : Optional[Any] = F'''{sampling_rate}''' lowerCAmelCase__ : Any = '1' if format_for_conversion == "s16le": lowerCAmelCase__ : Dict = 2 elif format_for_conversion == "f32le": lowerCAmelCase__ : List[str] = 4 else: raise ValueError(F'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) lowerCAmelCase__ : Tuple = platform.system() if system == "Linux": lowerCAmelCase__ : str = 'alsa' lowerCAmelCase__ : str = 'default' elif system == "Darwin": lowerCAmelCase__ : Any = 'avfoundation' lowerCAmelCase__ : Tuple = ':0' elif system == "Windows": lowerCAmelCase__ : Any = 'dshow' lowerCAmelCase__ : int = 'default' lowerCAmelCase__ : Any = [ 'ffmpeg', '-f', format_, '-i', input_, '-ac', ac, '-ar', ar, '-f', format_for_conversion, '-fflags', 'nobuffer', '-hide_banner', '-loglevel', 'quiet', 'pipe:1', ] lowerCAmelCase__ : List[str] = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample lowerCAmelCase__ : str = _ffmpeg_stream(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) for item in iterator: yield item def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = "f32le" , ) -> str: if stream_chunk_s is not None: lowerCAmelCase__ : Union[str, Any] = stream_chunk_s else: lowerCAmelCase__ : Tuple = chunk_length_s lowerCAmelCase__ : Any = ffmpeg_microphone(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , format_for_conversion=SCREAMING_SNAKE_CASE_ ) if format_for_conversion == "s16le": lowerCAmelCase__ : Optional[Any] = np.intaa lowerCAmelCase__ : Optional[Any] = 2 elif format_for_conversion == "f32le": lowerCAmelCase__ : Optional[Any] = np.floataa lowerCAmelCase__ : Optional[Any] = 4 else: raise ValueError(F'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) if stride_length_s is None: lowerCAmelCase__ : Dict = chunk_length_s / 6 lowerCAmelCase__ : int = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(SCREAMING_SNAKE_CASE_ , (int, float) ): lowerCAmelCase__ : Dict = [stride_length_s, stride_length_s] lowerCAmelCase__ : Union[str, Any] = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample lowerCAmelCase__ : List[Any] = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample lowerCAmelCase__ : Any = datetime.datetime.now() lowerCAmelCase__ : Any = datetime.timedelta(seconds=SCREAMING_SNAKE_CASE_ ) for item in chunk_bytes_iter(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , stride=(stride_left, stride_right) , stream=SCREAMING_SNAKE_CASE_ ): # Put everything back in numpy scale lowerCAmelCase__ : Any = np.frombuffer(item['raw'] , dtype=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = ( item['stride'][0] // size_of_sample, item['stride'][1] // size_of_sample, ) lowerCAmelCase__ : Optional[int] = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = False ) -> Optional[int]: lowerCAmelCase__ : Union[str, Any] = b'' lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = stride if stride_left + stride_right >= chunk_len: raise ValueError( F'''Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}''' ) lowerCAmelCase__ : List[str] = 0 for raw in iterator: acc += raw if stream and len(SCREAMING_SNAKE_CASE_ ) < chunk_len: lowerCAmelCase__ : Tuple = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(SCREAMING_SNAKE_CASE_ ) >= chunk_len: # We are flushing the accumulator lowerCAmelCase__ : Dict = (_stride_left, stride_right) lowerCAmelCase__ : Any = {'raw': acc[:chunk_len], 'stride': stride} if stream: lowerCAmelCase__ : Optional[int] = False yield item lowerCAmelCase__ : Optional[int] = stride_left lowerCAmelCase__ : Optional[int] = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(SCREAMING_SNAKE_CASE_ ) > stride_left: lowerCAmelCase__ : Tuple = {'raw': acc, 'stride': (_stride_left, 0)} if stream: lowerCAmelCase__ : Any = False yield item def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> str: lowerCAmelCase__ : int = 2**24 # 16Mo try: with subprocess.Popen(SCREAMING_SNAKE_CASE_ , stdout=subprocess.PIPE , bufsize=SCREAMING_SNAKE_CASE_ ) as ffmpeg_process: while True: lowerCAmelCase__ : List[str] = ffmpeg_process.stdout.read(SCREAMING_SNAKE_CASE_ ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError('ffmpeg was not found but is required to stream audio files from filename' ) from error
212
1
from __future__ import annotations import string from itertools import cycle, product from pathlib import Path _SCREAMING_SNAKE_CASE = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) _SCREAMING_SNAKE_CASE = [ord(letter) for letter in string.ascii_lowercase] _SCREAMING_SNAKE_CASE = {ord(char) for char in VALID_CHARS} _SCREAMING_SNAKE_CASE = ["the", "be", "to", "of", "and", "in", "that", "have"] def SCREAMING_SNAKE_CASE__ ( __a , __a ): snake_case_ : Dict = '' snake_case_ : Optional[int] = 42 snake_case_ : Union[str, Any] = 42 snake_case_ : Optional[int] = 42 for keychar, cipherchar in zip(cycle(__a ) , __a ): snake_case_ : Dict = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(__a ) return decoded def SCREAMING_SNAKE_CASE__ ( __a ): snake_case_ : Dict = [] for key in product(__a , repeat=3 ): snake_case_ : str = try_key(__a , __a ) if encoded is not None: possibles.append(__a ) return possibles def SCREAMING_SNAKE_CASE__ ( __a , __a ): return [possible for possible in possibles if common_word in possible.lower()] def SCREAMING_SNAKE_CASE__ ( __a = "p059_cipher.txt" ): snake_case_ : int = 42 snake_case_ : str = 42 snake_case_ : Dict = 42 snake_case_ : Optional[Any] = 42 snake_case_ : Optional[Any] = Path(__a ).parent.joinpath(__a ).read_text(encoding='utf-8' ) snake_case_ : Dict = [int(__a ) for number in data.strip().split(',' )] snake_case_ : str = filter_valid_chars(__a ) for common_word in COMMON_WORDS: snake_case_ : Tuple = filter_common_word(__a , __a ) if len(__a ) == 1: break snake_case_ : Any = possibles[0] return sum(ord(__a ) for char in decoded_text ) if __name__ == "__main__": print(F'''{solution() = }''')
364
import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() def SCREAMING_SNAKE_CASE__ ( __a , __a , __a , __a , __a ): # Initialise PyTorch model. # If you want to convert a checkpoint that uses absolute position embeddings, make sure to set reset_position_index_per_cell of # TapasConfig to False. # initialize configuration from json file snake_case_ : Optional[int] = TapasConfig.from_json_file(__a ) # set absolute/relative position embeddings parameter snake_case_ : List[Any] = reset_position_index_per_cell # set remaining parameters of TapasConfig as well as the model based on the task if task == "SQA": snake_case_ : int = TapasForQuestionAnswering(config=__a ) elif task == "WTQ": # run_task_main.py hparams snake_case_ : Optional[int] = 4 snake_case_ : List[str] = True # hparam_utils.py hparams snake_case_ : Optional[Any] = 0.664694 snake_case_ : Dict = 0.207951 snake_case_ : Tuple = 0.121194 snake_case_ : Dict = True snake_case_ : int = True snake_case_ : int = False snake_case_ : str = 0.0352513 snake_case_ : int = TapasForQuestionAnswering(config=__a ) elif task == "WIKISQL_SUPERVISED": # run_task_main.py hparams snake_case_ : int = 4 snake_case_ : Optional[int] = False # hparam_utils.py hparams snake_case_ : str = 36.4519 snake_case_ : Optional[Any] = 0.903421 snake_case_ : List[Any] = 222.088 snake_case_ : Optional[int] = True snake_case_ : Optional[Any] = True snake_case_ : str = True snake_case_ : int = 0.763141 snake_case_ : str = TapasForQuestionAnswering(config=__a ) elif task == "TABFACT": snake_case_ : List[Any] = TapasForSequenceClassification(config=__a ) elif task == "MLM": snake_case_ : Optional[int] = TapasForMaskedLM(config=__a ) elif task == "INTERMEDIATE_PRETRAINING": snake_case_ : Tuple = TapasModel(config=__a ) else: raise ValueError(f"""Task {task} not supported.""" ) print(f"""Building PyTorch model from configuration: {config}""" ) # Load weights from tf checkpoint load_tf_weights_in_tapas(__a , __a , __a ) # Save pytorch-model (weights and configuration) print(f"""Save PyTorch model to {pytorch_dump_path}""" ) model.save_pretrained(__a ) # Save tokenizer files print(f"""Save tokenizer files to {pytorch_dump_path}""" ) snake_case_ : str = TapasTokenizer(vocab_file=tf_checkpoint_path[:-10] + 'vocab.txt' , model_max_length=5_12 ) tokenizer.save_pretrained(__a ) print('Used relative position embeddings:' , model.config.reset_position_index_per_cell ) if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( """--task""", default="""SQA""", type=str, help="""Model task for which to convert a checkpoint. Defaults to SQA.""" ) parser.add_argument( """--reset_position_index_per_cell""", default=False, action="""store_true""", help="""Whether to use relative position embeddings or not. Defaults to True.""", ) parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--tapas_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained TAPAS model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) _SCREAMING_SNAKE_CASE = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.task, args.reset_position_index_per_cell, args.tf_checkpoint_path, args.tapas_config_file, args.pytorch_dump_path, )
88
0
"""simple docstring""" from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __UpperCamelCase : Optional[Any] = {"configuration_mra": ["MRA_PRETRAINED_CONFIG_ARCHIVE_MAP", "MraConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : Any = [ "MRA_PRETRAINED_MODEL_ARCHIVE_LIST", "MraForMaskedLM", "MraForMultipleChoice", "MraForQuestionAnswering", "MraForSequenceClassification", "MraForTokenClassification", "MraLayer", "MraModel", "MraPreTrainedModel", ] if TYPE_CHECKING: from .configuration_mra import MRA_PRETRAINED_CONFIG_ARCHIVE_MAP, MraConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mra import ( MRA_PRETRAINED_MODEL_ARCHIVE_LIST, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraLayer, MraModel, MraPreTrainedModel, ) else: import sys __UpperCamelCase : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
106
"""simple docstring""" from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import ( BaseOutput, OptionalDependencyNotAvailable, is_flax_available, is_k_diffusion_available, is_k_diffusion_version, is_onnx_available, is_torch_available, is_transformers_available, is_transformers_version, ) @dataclass class UpperCAmelCase_ ( _a): lowerCamelCase__ : Union[List[PIL.Image.Image], np.ndarray] lowerCamelCase__ : Optional[List[bool]] try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_cycle_diffusion import CycleDiffusionPipeline from .pipeline_stable_diffusion import StableDiffusionPipeline from .pipeline_stable_diffusion_attend_and_excite import StableDiffusionAttendAndExcitePipeline from .pipeline_stable_diffusion_imgaimg import StableDiffusionImgaImgPipeline from .pipeline_stable_diffusion_inpaint import StableDiffusionInpaintPipeline from .pipeline_stable_diffusion_inpaint_legacy import StableDiffusionInpaintPipelineLegacy from .pipeline_stable_diffusion_instruct_pixapix import StableDiffusionInstructPixaPixPipeline from .pipeline_stable_diffusion_latent_upscale import StableDiffusionLatentUpscalePipeline from .pipeline_stable_diffusion_ldmad import StableDiffusionLDMaDPipeline from .pipeline_stable_diffusion_model_editing import StableDiffusionModelEditingPipeline from .pipeline_stable_diffusion_panorama import StableDiffusionPanoramaPipeline from .pipeline_stable_diffusion_paradigms import StableDiffusionParadigmsPipeline from .pipeline_stable_diffusion_sag import StableDiffusionSAGPipeline from .pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from .pipeline_stable_unclip import StableUnCLIPPipeline from .pipeline_stable_unclip_imgaimg import StableUnCLIPImgaImgPipeline from .safety_checker import StableDiffusionSafetyChecker from .stable_unclip_image_normalizer import StableUnCLIPImageNormalizer try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.25.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import StableDiffusionImageVariationPipeline else: from .pipeline_stable_diffusion_image_variation import StableDiffusionImageVariationPipeline try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.26.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( StableDiffusionDepthaImgPipeline, StableDiffusionDiffEditPipeline, StableDiffusionPixaPixZeroPipeline, ) else: from .pipeline_stable_diffusion_depthaimg import StableDiffusionDepthaImgPipeline from .pipeline_stable_diffusion_diffedit import StableDiffusionDiffEditPipeline from .pipeline_stable_diffusion_pixapix_zero import StableDiffusionPixaPixZeroPipeline try: if not ( is_torch_available() and is_transformers_available() and is_k_diffusion_available() and is_k_diffusion_version(">=", "0.0.12") ): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_and_k_diffusion_objects import * # noqa F403 else: from .pipeline_stable_diffusion_k_diffusion import StableDiffusionKDiffusionPipeline try: if not (is_transformers_available() and is_onnx_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_onnx_objects import * # noqa F403 else: from .pipeline_onnx_stable_diffusion import OnnxStableDiffusionPipeline, StableDiffusionOnnxPipeline from .pipeline_onnx_stable_diffusion_imgaimg import OnnxStableDiffusionImgaImgPipeline from .pipeline_onnx_stable_diffusion_inpaint import OnnxStableDiffusionInpaintPipeline from .pipeline_onnx_stable_diffusion_inpaint_legacy import OnnxStableDiffusionInpaintPipelineLegacy from .pipeline_onnx_stable_diffusion_upscale import OnnxStableDiffusionUpscalePipeline if is_transformers_available() and is_flax_available(): import flax @flax.struct.dataclass class UpperCAmelCase_ ( _a): lowerCamelCase__ : np.ndarray lowerCamelCase__ : List[bool] from ...schedulers.scheduling_pndm_flax import PNDMSchedulerState from .pipeline_flax_stable_diffusion import FlaxStableDiffusionPipeline from .pipeline_flax_stable_diffusion_imgaimg import FlaxStableDiffusionImgaImgPipeline from .pipeline_flax_stable_diffusion_inpaint import FlaxStableDiffusionInpaintPipeline from .safety_checker_flax import FlaxStableDiffusionSafetyChecker
77
0
from jiwer import compute_measures import datasets SCREAMING_SNAKE_CASE_:str = """\ @inproceedings{inproceedings, author = {Morris, Andrew and Maier, Viktoria and Green, Phil}, year = {2004}, month = {01}, pages = {}, title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.} } """ SCREAMING_SNAKE_CASE_:Union[str, Any] = """\ Word error rate (WER) is a common metric of the performance of an automatic speech recognition system. The general difficulty of measuring performance lies in the fact that the recognized word sequence can have a different length from the reference word sequence (supposedly the correct one). The WER is derived from the Levenshtein distance, working at the word level instead of the phoneme level. The WER is a valuable tool for comparing different systems as well as for evaluating improvements within one system. This kind of measurement, however, provides no details on the nature of translation errors and further work is therefore required to identify the main source(s) of error and to focus any research effort. This problem is solved by first aligning the recognized word sequence with the reference (spoken) word sequence using dynamic string alignment. Examination of this issue is seen through a theory called the power law that states the correlation between perplexity and word error rate. Word error rate can then be computed as: WER = (S + D + I) / N = (S + D + I) / (S + D + C) where S is the number of substitutions, D is the number of deletions, I is the number of insertions, C is the number of correct words, N is the number of words in the reference (N=S+D+C). This value indicates the average number of errors per reference word. The lower the value, the better the performance of the ASR system with a WER of 0 being a perfect score. """ SCREAMING_SNAKE_CASE_:List[Any] = """ Compute WER score of transcribed segments against references. Args: references: List of references for each speech input. predictions: List of transcriptions to score. concatenate_texts (bool, default=False): Whether to concatenate all input texts or compute WER iteratively. Returns: (float): the word error rate Examples: >>> predictions = [\"this is the prediction\", \"there is an other sample\"] >>> references = [\"this is the reference\", \"there is another one\"] >>> wer = datasets.load_metric(\"wer\") >>> wer_score = wer.compute(predictions=predictions, references=references) >>> print(wer_score) 0.5 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class SCREAMING_SNAKE_CASE__ ( datasets.Metric ): '''simple docstring''' def _lowerCAmelCase ( self ): return datasets.MetricInfo( description=_DESCRIPTION, citation=_CITATION, inputs_description=_KWARGS_DESCRIPTION, features=datasets.Features( { """predictions""": datasets.Value("""string""", id="""sequence""" ), """references""": datasets.Value("""string""", id="""sequence""" ), } ), codebase_urls=["""https://github.com/jitsi/jiwer/"""], reference_urls=[ """https://en.wikipedia.org/wiki/Word_error_rate""", ], ) def _lowerCAmelCase ( self, lowerCamelCase__=None, lowerCamelCase__=None, lowerCamelCase__=False ): if concatenate_texts: return compute_measures(lowerCamelCase__, lowerCamelCase__ )["wer"] else: A : str = 0 A : int = 0 for prediction, reference in zip(lowerCamelCase__, lowerCamelCase__ ): A : Tuple = compute_measures(lowerCamelCase__, lowerCamelCase__ ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
115
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available SCREAMING_SNAKE_CASE_:Any = { """configuration_mobilenet_v2""": [ """MOBILENET_V2_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MobileNetV2Config""", """MobileNetV2OnnxConfig""", ], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_:List[Any] = ["""MobileNetV2FeatureExtractor"""] SCREAMING_SNAKE_CASE_:Tuple = ["""MobileNetV2ImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_:Any = [ """MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST""", """MobileNetV2ForImageClassification""", """MobileNetV2ForSemanticSegmentation""", """MobileNetV2Model""", """MobileNetV2PreTrainedModel""", """load_tf_weights_in_mobilenet_v2""", ] if TYPE_CHECKING: from .configuration_mobilenet_va import ( MOBILENET_V2_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileNetVaConfig, MobileNetVaOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_mobilenet_va import MobileNetVaFeatureExtractor from .image_processing_mobilenet_va import MobileNetVaImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilenet_va import ( MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST, MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation, MobileNetVaModel, MobileNetVaPreTrainedModel, load_tf_weights_in_mobilenet_va, ) else: import sys SCREAMING_SNAKE_CASE_:Optional[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
115
1
import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __snake_case ( __lowerCamelCase , unittest.TestCase ): '''simple docstring''' lowerCAmelCase__ = CodeGenTokenizer lowerCAmelCase__ = CodeGenTokenizerFast lowerCAmelCase__ = True lowerCAmelCase__ = {"""add_prefix_space""": True} lowerCAmelCase__ = False def UpperCAmelCase__ ( self : Optional[int] ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __snake_case: List[Any] = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", """<|endoftext|>""", ] __snake_case: Tuple = dict(zip(A , range(len(A ) ) ) ) __snake_case: Union[str, Any] = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] __snake_case: List[str] = {"""unk_token""": """<unk>"""} __snake_case: List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) __snake_case: Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(A ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(A ) ) def UpperCAmelCase__ ( self : Tuple , **A : Optional[int] ): kwargs.update(self.special_tokens_map ) return CodeGenTokenizer.from_pretrained(self.tmpdirname , **A ) def UpperCAmelCase__ ( self : Dict , **A : int ): kwargs.update(self.special_tokens_map ) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname , **A ) def UpperCAmelCase__ ( self : Dict , A : List[Any] ): __snake_case: int = """lower newer""" __snake_case: List[str] = """lower newer""" return input_text, output_text def UpperCAmelCase__ ( self : Tuple ): __snake_case: Union[str, Any] = CodeGenTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) __snake_case: int = """lower newer""" __snake_case: List[str] = ["""\u0120low""", """er""", """\u0120""", """n""", """e""", """w""", """er"""] __snake_case: Optional[int] = tokenizer.tokenize(A , add_prefix_space=A ) self.assertListEqual(A , A ) __snake_case: Tuple = tokens + [tokenizer.unk_token] __snake_case: Tuple = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(A ) , A ) def UpperCAmelCase__ ( self : int ): if not self.test_rust_tokenizer: return __snake_case: List[str] = self.get_tokenizer() __snake_case: List[Any] = self.get_rust_tokenizer(add_prefix_space=A ) __snake_case: int = """lower newer""" # Testing tokenization __snake_case: Dict = tokenizer.tokenize(A , add_prefix_space=A ) __snake_case: int = rust_tokenizer.tokenize(A ) self.assertListEqual(A , A ) # Testing conversion to ids without special tokens __snake_case: List[Any] = tokenizer.encode(A , add_special_tokens=A , add_prefix_space=A ) __snake_case: Union[str, Any] = rust_tokenizer.encode(A , add_special_tokens=A ) self.assertListEqual(A , A ) # Testing conversion to ids with special tokens __snake_case: int = self.get_rust_tokenizer(add_prefix_space=A ) __snake_case: Optional[Any] = tokenizer.encode(A , add_prefix_space=A ) __snake_case: Optional[Any] = rust_tokenizer.encode(A ) self.assertListEqual(A , A ) # Testing the unknown token __snake_case: str = tokens + [rust_tokenizer.unk_token] __snake_case: str = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(A ) , A ) def UpperCAmelCase__ ( self : Tuple , *A : Tuple , **A : List[str] ): # It's very difficult to mix/test pretokenization with byte-level # And get both CodeGen and Roberta to work at the same time (mostly an issue of adding a space before the string) pass def UpperCAmelCase__ ( self : Dict , A : Union[str, Any]=15 ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): __snake_case: Optional[Any] = self.rust_tokenizer_class.from_pretrained(A , **A ) # Simple input __snake_case: List[str] = """This is a simple input""" __snake_case: Optional[Any] = ["""This is a simple input 1""", """This is a simple input 2"""] __snake_case: List[str] = ("""This is a simple input""", """This is a pair""") __snake_case: Dict = [ ("""This is a simple input 1""", """This is a simple input 2"""), ("""This is a simple pair 1""", """This is a simple pair 2"""), ] # Simple input tests self.assertRaises(A , tokenizer_r.encode , A , max_length=A , padding="""max_length""" ) # Simple input self.assertRaises(A , tokenizer_r.encode_plus , A , max_length=A , padding="""max_length""" ) # Simple input self.assertRaises( A , tokenizer_r.batch_encode_plus , A , max_length=A , padding="""max_length""" , ) # Pair input self.assertRaises(A , tokenizer_r.encode , A , max_length=A , padding="""max_length""" ) # Pair input self.assertRaises(A , tokenizer_r.encode_plus , A , max_length=A , padding="""max_length""" ) # Pair input self.assertRaises( A , tokenizer_r.batch_encode_plus , A , max_length=A , padding="""max_length""" , ) def UpperCAmelCase__ ( self : Union[str, Any] ): __snake_case: Optional[Any] = CodeGenTokenizer.from_pretrained(self.tmpdirname , pad_token="""<pad>""" ) # Simple input __snake_case: Dict = """This is a simple input""" __snake_case: int = ["""This is a simple input looooooooong""", """This is a simple input"""] __snake_case: str = ("""This is a simple input""", """This is a pair""") __snake_case: Any = [ ("""This is a simple input loooooong""", """This is a simple input"""), ("""This is a simple pair loooooong""", """This is a simple pair"""), ] __snake_case: str = tokenizer.pad_token_id __snake_case: List[Any] = tokenizer(A , padding="""max_length""" , max_length=30 , return_tensors="""np""" ) __snake_case: Union[str, Any] = tokenizer(A , padding=A , truncate=A , return_tensors="""np""" ) __snake_case: Union[str, Any] = tokenizer(*A , padding="""max_length""" , max_length=60 , return_tensors="""np""" ) __snake_case: int = tokenizer(A , padding=A , truncate=A , return_tensors="""np""" ) # s # test single string max_length padding self.assertEqual(out_s["""input_ids"""].shape[-1] , 30 ) self.assertTrue(pad_token_id in out_s["""input_ids"""] ) self.assertTrue(0 in out_s["""attention_mask"""] ) # s2 # test automatic padding self.assertEqual(out_sa["""input_ids"""].shape[-1] , 33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa["""input_ids"""][0] ) self.assertFalse(0 in out_sa["""attention_mask"""][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa["""input_ids"""][1] ) self.assertTrue(0 in out_sa["""attention_mask"""][1] ) # p # test single pair max_length padding self.assertEqual(out_p["""input_ids"""].shape[-1] , 60 ) self.assertTrue(pad_token_id in out_p["""input_ids"""] ) self.assertTrue(0 in out_p["""attention_mask"""] ) # p2 # test automatic padding pair self.assertEqual(out_pa["""input_ids"""].shape[-1] , 52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa["""input_ids"""][0] ) self.assertFalse(0 in out_pa["""attention_mask"""][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa["""input_ids"""][1] ) self.assertTrue(0 in out_pa["""attention_mask"""][1] ) def UpperCAmelCase__ ( self : Optional[int] ): __snake_case: Dict = """$$$""" __snake_case: List[str] = CodeGenTokenizer.from_pretrained(self.tmpdirname , bos_token=A , add_bos_token=A ) __snake_case: Any = """This is a simple input""" __snake_case: int = ["""This is a simple input 1""", """This is a simple input 2"""] __snake_case: Tuple = tokenizer.bos_token_id __snake_case: List[str] = tokenizer(A ) __snake_case: Optional[Any] = tokenizer(A ) self.assertEqual(out_s.input_ids[0] , A ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) __snake_case: Union[str, Any] = tokenizer.decode(out_s.input_ids ) __snake_case: int = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , A ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) @slow def UpperCAmelCase__ ( self : List[str] ): __snake_case: Optional[int] = CodeGenTokenizer.from_pretrained("""Salesforce/codegen-350M-mono""" ) __snake_case: List[str] = """\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#""" __snake_case: List[str] = """\nif len_a > len_b: result = a\nelse: result = b""" __snake_case: List[str] = tokenizer.encode(A ) __snake_case: Union[str, Any] = ["""^#""", re.escape("""<|endoftext|>""" ), """^'''""", """^\"\"\"""", """\n\n\n"""] __snake_case: str = tokenizer.decode(A , truncate_before_pattern=A ) self.assertEqual(A , A ) def UpperCAmelCase__ ( self : int ): pass
111
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available __UpperCAmelCase : Union[str, Any] = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase : str = ["BartphoTokenizer"] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bartpho import BartphoTokenizer else: import sys __UpperCAmelCase : Optional[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
111
1
"""simple docstring""" import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''spiece.model'''} lowerCAmelCase__ = { '''vocab_file''': { '''AI-Sweden/gpt-sw3-126m''': '''https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model''', '''AI-Sweden/gpt-sw3-350m''': '''https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model''', '''AI-Sweden/gpt-sw3-1.6b''': '''https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model''', '''AI-Sweden/gpt-sw3-6.7b''': '''https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model''', '''AI-Sweden/gpt-sw3-20b''': '''https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model''', } } lowerCAmelCase__ = { '''AI-Sweden/gpt-sw3-126m''': 2_048, '''AI-Sweden/gpt-sw3-350m''': 2_048, '''AI-Sweden/gpt-sw3-1.6b''': 2_048, '''AI-Sweden/gpt-sw3-6.7b''': 2_048, '''AI-Sweden/gpt-sw3-20b''': 2_048, } class SCREAMING_SNAKE_CASE__ ( A__ ): """simple docstring""" a : str =VOCAB_FILES_NAMES a : List[str] =PRETRAINED_VOCAB_FILES_MAP a : int =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a : str =["input_ids", "attention_mask"] def __init__( self , snake_case__ , snake_case__=False , snake_case__=False , snake_case__=False , snake_case__=None , snake_case__=None , snake_case__=None , snake_case__=None , snake_case__ = None , **snake_case__ , ): """simple docstring""" lowerCAmelCase : Optional[Any] = {} if sp_model_kwargs is None else sp_model_kwargs lowerCAmelCase : Any = kwargs.get("name_or_path" ) if name_or_path is None: logger.warning( "name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b," " you are testing the model, this can safely be ignored" ) lowerCAmelCase : str = """None""" # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing lowerCAmelCase : Dict = """<|endoftext|>""" if eos_token is None else eos_token lowerCAmelCase : Any = """<unk>""" if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: lowerCAmelCase : Dict = unk_token if pad_token is None else pad_token lowerCAmelCase : List[Any] = eos_token if bos_token is None else bos_token else: lowerCAmelCase : Any = """<pad>""" if pad_token is None else pad_token lowerCAmelCase : Any = """<s>""" if bos_token is None else bos_token super().__init__( do_lower_case=__A , remove_space=__A , keep_accents=__A , bos_token=__A , eos_token=__A , unk_token=__A , pad_token=__A , sp_model_kwargs=self.sp_model_kwargs , **__A , ) lowerCAmelCase : List[Any] = do_lower_case lowerCAmelCase : Dict = remove_space lowerCAmelCase : Optional[int] = keep_accents lowerCAmelCase : int = vocab_file lowerCAmelCase : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__A ) # Used for whitespace normalization in input texts # fmt : off lowerCAmelCase : List[str] = {""" """, """ """, """ """, """ """, """ """, """ """, """ """, """ """, """ """, """ """, """""", """„"""} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing lowerCAmelCase : int = re.compile( f"""[{"".join(map(__A , list(range(0 , 9 ) ) + list(range(11 , 32 ) ) + list(range(127 , 160 ) ) + [160, 173, 8_203] ) )}]""" ) def __getstate__( self ): """simple docstring""" lowerCAmelCase : Dict = self.__dict__.copy() lowerCAmelCase : str = None return state def __setstate__( self , snake_case__ ): """simple docstring""" lowerCAmelCase : int = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): lowerCAmelCase : Optional[Any] = {} lowerCAmelCase : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def lowercase__ ( self ): """simple docstring""" return len(self.sp_model ) def lowercase__ ( self , snake_case__ ): """simple docstring""" lowerCAmelCase : Any = self.non_printing_characters_re.sub("" , __A ) # Normalize whitespaces lowerCAmelCase : str = """""".join([char if char not in self.whitespaces else " " for char in text] ) # NFC Unicode normalization lowerCAmelCase : Union[str, Any] = unicodedata.normalize("NFC" , __A ) return text def lowercase__ ( self , snake_case__ , **snake_case__ ): """simple docstring""" lowerCAmelCase : Dict = self.preprocess_text(__A ) return self.sp_model.encode(__A , out_type=__A ) def lowercase__ ( self , snake_case__ ): """simple docstring""" return self.sp_model.PieceToId(__A ) def lowercase__ ( self , snake_case__ ): """simple docstring""" return self.sp_model.IdToPiece(__A ) @staticmethod def lowercase__ ( snake_case__ ): """simple docstring""" return out_string def lowercase__ ( self , snake_case__ ): """simple docstring""" lowerCAmelCase : int = [] lowerCAmelCase : int = """""" lowerCAmelCase : List[Any] = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__A ) + token lowerCAmelCase : Tuple = True lowerCAmelCase : Optional[int] = [] else: current_sub_tokens.append(__A ) lowerCAmelCase : str = False out_string += self.sp_model.decode(__A ) return out_string def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : int = {self.convert_ids_to_tokens(__A ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def lowercase__ ( self , snake_case__ , snake_case__ = None ): """simple docstring""" if not os.path.isdir(__A ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowerCAmelCase : Optional[int] = os.path.join( __A , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__A ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __A ) elif not os.path.isfile(self.vocab_file ): with open(__A , "wb" ) as fi: lowerCAmelCase : str = self.sp_model.serialized_model_proto() fi.write(__A ) return (out_vocab_file,) def lowercase__ ( self , snake_case__ , snake_case__ = False ): """simple docstring""" if isinstance(__A , __A ): lowerCAmelCase : Dict = self.preprocess_text(__A ) lowerCAmelCase : int = self.sp_model.encode(__A ) else: lowerCAmelCase : List[str] = [self.preprocess_text(__A ) for t in text] lowerCAmelCase : List[Any] = self.sp_model.encode(__A ) if return_tensors is True or return_tensors == "pt": lowerCAmelCase : Dict = torch.tensor(__A ) return token_ids def lowercase__ ( self , snake_case__ ): """simple docstring""" return self.sp_model.decode(__A ) def lowercase__ ( self , snake_case__ ): """simple docstring""" lowerCAmelCase : str = [f"""User: {text}""" if is_user else f"""Bot: {text}""" for is_user, text in conversation.iter_texts()] lowerCAmelCase : Optional[Any] = ( f"""{self.eos_token}{self.bos_token}""" + f"""{self.bos_token}""".join(__A ) + f"""{self.bos_token}Bot:""" ) return self.encode(text=__A )
351
"""simple docstring""" def a__ ( SCREAMING_SNAKE_CASE : list[int] ): '''simple docstring''' lowerCAmelCase : str = len(SCREAMING_SNAKE_CASE ) for i in range(SCREAMING_SNAKE_CASE ): for j in range(i + 1 , SCREAMING_SNAKE_CASE ): if numbers[j] < numbers[i]: lowerCAmelCase , lowerCAmelCase : Any = numbers[j], numbers[i] return numbers if __name__ == "__main__": lowerCAmelCase__ = input('''Enter numbers separated by a comma:\n''').strip() lowerCAmelCase__ = [int(item) for item in user_input.split(''',''')] print(exchange_sort(unsorted))
133
0
'''simple docstring''' from typing import Optional, Tuple import jax import jax.numpy as jnp from flax import linen as nn from flax.core.frozen_dict import FrozenDict from transformers import CLIPConfig, FlaxPreTrainedModel from transformers.models.clip.modeling_flax_clip import FlaxCLIPVisionModule def a_ ( _lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase=1E-12 ) -> Union[str, Any]: __lowerCamelCase : Any = jnp.divide(emb_a.T ,jnp.clip(jnp.linalg.norm(_lowerCAmelCase ,axis=1 ) ,a_min=_lowerCAmelCase ) ).T __lowerCamelCase : str = jnp.divide(emb_a.T ,jnp.clip(jnp.linalg.norm(_lowerCAmelCase ,axis=1 ) ,a_min=_lowerCAmelCase ) ).T return jnp.matmul(_lowerCAmelCase ,norm_emb_a.T ) class lowerCamelCase_ ( nn.Module ): """simple docstring""" a_ =42 a_ =jnp.floataa def _lowercase ( self : Optional[int] ) -> List[str]: __lowerCamelCase : Tuple = FlaxCLIPVisionModule(self.config.vision_config ) __lowerCamelCase : Optional[Any] = nn.Dense(self.config.projection_dim , use_bias=__snake_case , dtype=self.dtype ) __lowerCamelCase : Any = self.param('concept_embeds' , jax.nn.initializers.ones , (17, self.config.projection_dim) ) __lowerCamelCase : List[str] = self.param( 'special_care_embeds' , jax.nn.initializers.ones , (3, self.config.projection_dim) ) __lowerCamelCase : Any = self.param('concept_embeds_weights' , jax.nn.initializers.ones , (17,) ) __lowerCamelCase : Tuple = self.param('special_care_embeds_weights' , jax.nn.initializers.ones , (3,) ) def __call__( self : Optional[Any] , _a : Tuple ) -> int: __lowerCamelCase : List[Any] = self.vision_model(__snake_case )[1] __lowerCamelCase : List[Any] = self.visual_projection(__snake_case ) __lowerCamelCase : List[str] = jax_cosine_distance(__snake_case , self.special_care_embeds ) __lowerCamelCase : Optional[Any] = jax_cosine_distance(__snake_case , self.concept_embeds ) # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign image inputs __lowerCamelCase : Any = 0.0 __lowerCamelCase : int = special_cos_dist - self.special_care_embeds_weights[None, :] + adjustment __lowerCamelCase : int = jnp.round(__snake_case , 3 ) __lowerCamelCase : Optional[int] = jnp.any(special_scores > 0 , axis=1 , keepdims=__snake_case ) # Use a lower threshold if an image has any special care concept __lowerCamelCase : Optional[int] = is_special_care * 0.01 __lowerCamelCase : List[str] = cos_dist - self.concept_embeds_weights[None, :] + special_adjustment __lowerCamelCase : str = jnp.round(__snake_case , 3 ) __lowerCamelCase : Union[str, Any] = jnp.any(concept_scores > 0 , axis=1 ) return has_nsfw_concepts class lowerCamelCase_ ( snake_case_ ): """simple docstring""" a_ =CLIPConfig a_ ='''clip_input''' a_ =FlaxStableDiffusionSafetyCheckerModule def __init__( self : List[str] , _a : CLIPConfig , _a : Optional[Tuple] = None , _a : int = 0 , _a : jnp.dtype = jnp.floataa , _a : bool = True , **_a : str , ) -> List[str]: if input_shape is None: __lowerCamelCase : Dict = (1, 224, 224, 3) __lowerCamelCase : Optional[Any] = self.module_class(config=__snake_case , dtype=__snake_case , **__snake_case ) super().__init__(__snake_case , __snake_case , input_shape=__snake_case , seed=__snake_case , dtype=__snake_case , _do_init=_do_init ) def _lowercase ( self : str , _a : jax.random.KeyArray , _a : Tuple , _a : FrozenDict = None ) -> FrozenDict: # init input tensor __lowerCamelCase : str = jax.random.normal(__snake_case , __snake_case ) __lowerCamelCase ,__lowerCamelCase : str = jax.random.split(__snake_case ) __lowerCamelCase : Union[str, Any] = {'params': params_rng, 'dropout': dropout_rng} __lowerCamelCase : Dict = self.module.init(__snake_case , __snake_case )['params'] return random_params def __call__( self : Any , _a : Optional[int] , _a : dict = None , ) -> int: __lowerCamelCase : Dict = jnp.transpose(__snake_case , (0, 2, 3, 1) ) return self.module.apply( {'params': params or self.params} , jnp.array(__snake_case , dtype=jnp.floataa ) , rngs={} , )
208
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from transformers.utils import is_vision_available from transformers.utils.generic import TensorType from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import logging if is_vision_available(): import PIL A__ : List[str] =logging.get_logger(__name__) def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" if isinstance(lowerCAmelCase , (list, tuple) ) and isinstance(videos[0] , (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(lowerCAmelCase , (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(lowerCAmelCase ): return [[videos]] raise ValueError(f"Could not make batched video from {videos}" ) class UpperCAmelCase ( snake_case_ ): _lowercase: Any = ['''pixel_values'''] def __init__( self : Tuple , __snake_case : bool = True , __snake_case : Dict[str, int] = None , __snake_case : PILImageResampling = PILImageResampling.BILINEAR , __snake_case : bool = True , __snake_case : Dict[str, int] = None , __snake_case : bool = True , __snake_case : Union[int, float] = 1 / 2_55 , __snake_case : bool = True , __snake_case : bool = True , __snake_case : Optional[Union[float, List[float]]] = None , __snake_case : Optional[Union[float, List[float]]] = None , **__snake_case : str , ) -> None: super().__init__(**__snake_case ) _lowerCAmelCase = size if size is not None else {"""shortest_edge""": 2_56} _lowerCAmelCase = get_size_dict(__snake_case , default_to_square=__snake_case ) _lowerCAmelCase = crop_size if crop_size is not None else {"""height""": 2_24, """width""": 2_24} _lowerCAmelCase = get_size_dict(__snake_case , param_name="""crop_size""" ) _lowerCAmelCase = do_resize _lowerCAmelCase = size _lowerCAmelCase = do_center_crop _lowerCAmelCase = crop_size _lowerCAmelCase = resample _lowerCAmelCase = do_rescale _lowerCAmelCase = rescale_factor _lowerCAmelCase = offset _lowerCAmelCase = do_normalize _lowerCAmelCase = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _lowerCAmelCase = image_std if image_std is not None else IMAGENET_STANDARD_STD def lowercase__ ( self : int , __snake_case : np.ndarray , __snake_case : Dict[str, int] , __snake_case : PILImageResampling = PILImageResampling.BILINEAR , __snake_case : Optional[Union[str, ChannelDimension]] = None , **__snake_case : Optional[Any] , ) -> np.ndarray: _lowerCAmelCase = get_size_dict(__snake_case , default_to_square=__snake_case ) if "shortest_edge" in size: _lowerCAmelCase = get_resize_output_image_size(__snake_case , size["""shortest_edge"""] , default_to_square=__snake_case ) elif "height" in size and "width" in size: _lowerCAmelCase = (size["""height"""], size["""width"""]) else: raise ValueError(f"Size must have 'height' and 'width' or 'shortest_edge' as keys. Got {size.keys()}" ) return resize(__snake_case , size=__snake_case , resample=__snake_case , data_format=__snake_case , **__snake_case ) def lowercase__ ( self : Union[str, Any] , __snake_case : np.ndarray , __snake_case : Dict[str, int] , __snake_case : Optional[Union[str, ChannelDimension]] = None , **__snake_case : List[Any] , ) -> np.ndarray: _lowerCAmelCase = get_size_dict(__snake_case ) if "height" not in size or "width" not in size: raise ValueError(f"Size must have 'height' and 'width' as keys. Got {size.keys()}" ) return center_crop(__snake_case , size=(size["""height"""], size["""width"""]) , data_format=__snake_case , **__snake_case ) def lowercase__ ( self : Union[str, Any] , __snake_case : np.ndarray , __snake_case : Union[int, float] , __snake_case : bool = True , __snake_case : Optional[Union[str, ChannelDimension]] = None , **__snake_case : Optional[Any] , ) -> Dict: _lowerCAmelCase = image.astype(np.floataa ) if offset: _lowerCAmelCase = image - (scale / 2) return rescale(__snake_case , scale=__snake_case , data_format=__snake_case , **__snake_case ) def lowercase__ ( self : Optional[int] , __snake_case : np.ndarray , __snake_case : Union[float, List[float]] , __snake_case : Union[float, List[float]] , __snake_case : Optional[Union[str, ChannelDimension]] = None , **__snake_case : Tuple , ) -> np.ndarray: return normalize(__snake_case , mean=__snake_case , std=__snake_case , data_format=__snake_case , **__snake_case ) def lowercase__ ( self : List[Any] , __snake_case : ImageInput , __snake_case : bool = None , __snake_case : Dict[str, int] = None , __snake_case : PILImageResampling = None , __snake_case : bool = None , __snake_case : Dict[str, int] = None , __snake_case : bool = None , __snake_case : float = None , __snake_case : bool = None , __snake_case : bool = None , __snake_case : Optional[Union[float, List[float]]] = None , __snake_case : Optional[Union[float, List[float]]] = None , __snake_case : Optional[ChannelDimension] = ChannelDimension.FIRST , ) -> np.ndarray: if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) if offset and not do_rescale: raise ValueError("""For offset, do_rescale must also be set to True.""" ) # All transformations expect numpy arrays. _lowerCAmelCase = to_numpy_array(__snake_case ) if do_resize: _lowerCAmelCase = self.resize(image=__snake_case , size=__snake_case , resample=__snake_case ) if do_center_crop: _lowerCAmelCase = self.center_crop(__snake_case , size=__snake_case ) if do_rescale: _lowerCAmelCase = self.rescale(image=__snake_case , scale=__snake_case , offset=__snake_case ) if do_normalize: _lowerCAmelCase = self.normalize(image=__snake_case , mean=__snake_case , std=__snake_case ) _lowerCAmelCase = to_channel_dimension_format(__snake_case , __snake_case ) return image def lowercase__ ( self : List[Any] , __snake_case : ImageInput , __snake_case : bool = None , __snake_case : Dict[str, int] = None , __snake_case : PILImageResampling = None , __snake_case : bool = None , __snake_case : Dict[str, int] = None , __snake_case : bool = None , __snake_case : float = None , __snake_case : bool = None , __snake_case : bool = None , __snake_case : Optional[Union[float, List[float]]] = None , __snake_case : Optional[Union[float, List[float]]] = None , __snake_case : Optional[Union[str, TensorType]] = None , __snake_case : ChannelDimension = ChannelDimension.FIRST , **__snake_case : List[str] , ) -> PIL.Image.Image: _lowerCAmelCase = do_resize if do_resize is not None else self.do_resize _lowerCAmelCase = resample if resample is not None else self.resample _lowerCAmelCase = do_center_crop if do_center_crop is not None else self.do_center_crop _lowerCAmelCase = do_rescale if do_rescale is not None else self.do_rescale _lowerCAmelCase = rescale_factor if rescale_factor is not None else self.rescale_factor _lowerCAmelCase = offset if offset is not None else self.offset _lowerCAmelCase = do_normalize if do_normalize is not None else self.do_normalize _lowerCAmelCase = image_mean if image_mean is not None else self.image_mean _lowerCAmelCase = image_std if image_std is not None else self.image_std _lowerCAmelCase = size if size is not None else self.size _lowerCAmelCase = get_size_dict(__snake_case , default_to_square=__snake_case ) _lowerCAmelCase = crop_size if crop_size is not None else self.crop_size _lowerCAmelCase = get_size_dict(__snake_case , param_name="""crop_size""" ) if not valid_images(__snake_case ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) _lowerCAmelCase = make_batched(__snake_case ) _lowerCAmelCase = [ [ self._preprocess_image( image=__snake_case , do_resize=__snake_case , size=__snake_case , resample=__snake_case , do_center_crop=__snake_case , crop_size=__snake_case , do_rescale=__snake_case , rescale_factor=__snake_case , offset=__snake_case , do_normalize=__snake_case , image_mean=__snake_case , image_std=__snake_case , data_format=__snake_case , ) for img in video ] for video in videos ] _lowerCAmelCase = {"""pixel_values""": videos} return BatchFeature(data=__snake_case , tensor_type=__snake_case )
70
0
"""simple docstring""" from typing import Any, Callable, Dict, List, Optional, Union import torch from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker _A = """CompVis/stable-diffusion-v1-1""" _A = """CompVis/stable-diffusion-v1-2""" _A = """CompVis/stable-diffusion-v1-3""" _A = """CompVis/stable-diffusion-v1-4""" class lowerCamelCase ( _a ): '''simple docstring''' def __init__(self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = True , ): """simple docstring""" super()._init_() UpperCAmelCase__ : int = StableDiffusionPipeline.from_pretrained(snake_case_ ) UpperCAmelCase__ : Optional[int] = StableDiffusionPipeline.from_pretrained(snake_case_ ) UpperCAmelCase__ : Dict = StableDiffusionPipeline.from_pretrained(snake_case_ ) UpperCAmelCase__ : int = StableDiffusionPipeline( vae=snake_case_ , text_encoder=snake_case_ , tokenizer=snake_case_ , unet=snake_case_ , scheduler=snake_case_ , safety_checker=snake_case_ , feature_extractor=snake_case_ , requires_safety_checker=snake_case_ , ) self.register_modules(pipelinea=self.pipea , pipelinea=self.pipea , pipelinea=self.pipea , pipelinea=self.pipea ) @property def _a (self ): """simple docstring""" return {k: getattr(self , snake_case_ ) for k in self.config.keys() if not k.startswith("""_""" )} def _a (self , _lowerCamelCase = "auto" ): """simple docstring""" if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory UpperCAmelCase__ : Dict = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(snake_case_ ) def _a (self ): """simple docstring""" self.enable_attention_slicing(snake_case_ ) @torch.no_grad() def _a (self , _lowerCamelCase , _lowerCamelCase = 512 , _lowerCamelCase = 512 , _lowerCamelCase = 50 , _lowerCamelCase = 7.5 , _lowerCamelCase = None , _lowerCamelCase = 1 , _lowerCamelCase = 0.0 , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = "pil" , _lowerCamelCase = True , _lowerCamelCase = None , _lowerCamelCase = 1 , **_lowerCamelCase , ): """simple docstring""" return self.pipea( prompt=snake_case_ , height=snake_case_ , width=snake_case_ , num_inference_steps=snake_case_ , guidance_scale=snake_case_ , negative_prompt=snake_case_ , num_images_per_prompt=snake_case_ , eta=snake_case_ , generator=snake_case_ , latents=snake_case_ , output_type=snake_case_ , return_dict=snake_case_ , callback=snake_case_ , callback_steps=snake_case_ , **snake_case_ , ) @torch.no_grad() def _a (self , _lowerCamelCase , _lowerCamelCase = 512 , _lowerCamelCase = 512 , _lowerCamelCase = 50 , _lowerCamelCase = 7.5 , _lowerCamelCase = None , _lowerCamelCase = 1 , _lowerCamelCase = 0.0 , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = "pil" , _lowerCamelCase = True , _lowerCamelCase = None , _lowerCamelCase = 1 , **_lowerCamelCase , ): """simple docstring""" return self.pipea( prompt=snake_case_ , height=snake_case_ , width=snake_case_ , num_inference_steps=snake_case_ , guidance_scale=snake_case_ , negative_prompt=snake_case_ , num_images_per_prompt=snake_case_ , eta=snake_case_ , generator=snake_case_ , latents=snake_case_ , output_type=snake_case_ , return_dict=snake_case_ , callback=snake_case_ , callback_steps=snake_case_ , **snake_case_ , ) @torch.no_grad() def _a (self , _lowerCamelCase , _lowerCamelCase = 512 , _lowerCamelCase = 512 , _lowerCamelCase = 50 , _lowerCamelCase = 7.5 , _lowerCamelCase = None , _lowerCamelCase = 1 , _lowerCamelCase = 0.0 , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = "pil" , _lowerCamelCase = True , _lowerCamelCase = None , _lowerCamelCase = 1 , **_lowerCamelCase , ): """simple docstring""" return self.pipea( prompt=snake_case_ , height=snake_case_ , width=snake_case_ , num_inference_steps=snake_case_ , guidance_scale=snake_case_ , negative_prompt=snake_case_ , num_images_per_prompt=snake_case_ , eta=snake_case_ , generator=snake_case_ , latents=snake_case_ , output_type=snake_case_ , return_dict=snake_case_ , callback=snake_case_ , callback_steps=snake_case_ , **snake_case_ , ) @torch.no_grad() def _a (self , _lowerCamelCase , _lowerCamelCase = 512 , _lowerCamelCase = 512 , _lowerCamelCase = 50 , _lowerCamelCase = 7.5 , _lowerCamelCase = None , _lowerCamelCase = 1 , _lowerCamelCase = 0.0 , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = "pil" , _lowerCamelCase = True , _lowerCamelCase = None , _lowerCamelCase = 1 , **_lowerCamelCase , ): """simple docstring""" return self.pipea( prompt=snake_case_ , height=snake_case_ , width=snake_case_ , num_inference_steps=snake_case_ , guidance_scale=snake_case_ , negative_prompt=snake_case_ , num_images_per_prompt=snake_case_ , eta=snake_case_ , generator=snake_case_ , latents=snake_case_ , output_type=snake_case_ , return_dict=snake_case_ , callback=snake_case_ , callback_steps=snake_case_ , **snake_case_ , ) @torch.no_grad() def _a (self , _lowerCamelCase , _lowerCamelCase = 512 , _lowerCamelCase = 512 , _lowerCamelCase = 50 , _lowerCamelCase = 7.5 , _lowerCamelCase = None , _lowerCamelCase = 1 , _lowerCamelCase = 0.0 , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = "pil" , _lowerCamelCase = True , _lowerCamelCase = None , _lowerCamelCase = 1 , **_lowerCamelCase , ): """simple docstring""" UpperCAmelCase__ : Union[str, Any] = """cuda""" if torch.cuda.is_available() else """cpu""" self.to(snake_case_ ) # Checks if the height and width are divisible by 8 or not if height % 8 != 0 or width % 8 != 0: raise ValueError(F"""`height` and `width` must be divisible by 8 but are {height} and {width}.""" ) # Get first result from Stable Diffusion Checkpoint v1.1 UpperCAmelCase__ : int = self.textaimg_sda_a( prompt=snake_case_ , height=snake_case_ , width=snake_case_ , num_inference_steps=snake_case_ , guidance_scale=snake_case_ , negative_prompt=snake_case_ , num_images_per_prompt=snake_case_ , eta=snake_case_ , generator=snake_case_ , latents=snake_case_ , output_type=snake_case_ , return_dict=snake_case_ , callback=snake_case_ , callback_steps=snake_case_ , **snake_case_ , ) # Get first result from Stable Diffusion Checkpoint v1.2 UpperCAmelCase__ : Any = self.textaimg_sda_a( prompt=snake_case_ , height=snake_case_ , width=snake_case_ , num_inference_steps=snake_case_ , guidance_scale=snake_case_ , negative_prompt=snake_case_ , num_images_per_prompt=snake_case_ , eta=snake_case_ , generator=snake_case_ , latents=snake_case_ , output_type=snake_case_ , return_dict=snake_case_ , callback=snake_case_ , callback_steps=snake_case_ , **snake_case_ , ) # Get first result from Stable Diffusion Checkpoint v1.3 UpperCAmelCase__ : Any = self.textaimg_sda_a( prompt=snake_case_ , height=snake_case_ , width=snake_case_ , num_inference_steps=snake_case_ , guidance_scale=snake_case_ , negative_prompt=snake_case_ , num_images_per_prompt=snake_case_ , eta=snake_case_ , generator=snake_case_ , latents=snake_case_ , output_type=snake_case_ , return_dict=snake_case_ , callback=snake_case_ , callback_steps=snake_case_ , **snake_case_ , ) # Get first result from Stable Diffusion Checkpoint v1.4 UpperCAmelCase__ : Any = self.textaimg_sda_a( prompt=snake_case_ , height=snake_case_ , width=snake_case_ , num_inference_steps=snake_case_ , guidance_scale=snake_case_ , negative_prompt=snake_case_ , num_images_per_prompt=snake_case_ , eta=snake_case_ , generator=snake_case_ , latents=snake_case_ , output_type=snake_case_ , return_dict=snake_case_ , callback=snake_case_ , callback_steps=snake_case_ , **snake_case_ , ) # Get all result images into a single list and pass it via StableDiffusionPipelineOutput for final result return StableDiffusionPipelineOutput([resa[0], resa[0], resa[0], resa[0]] )
371
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_fnet import FNetTokenizer else: _A = None _A = logging.get_logger(__name__) _A = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} _A = { """vocab_file""": { """google/fnet-base""": """https://huggingface.co/google/fnet-base/resolve/main/spiece.model""", """google/fnet-large""": """https://huggingface.co/google/fnet-large/resolve/main/spiece.model""", }, """tokenizer_file""": { """google/fnet-base""": """https://huggingface.co/google/fnet-base/resolve/main/tokenizer.json""", """google/fnet-large""": """https://huggingface.co/google/fnet-large/resolve/main/tokenizer.json""", }, } _A = { """google/fnet-base""": 5_12, """google/fnet-large""": 5_12, } _A = """▁""" class lowerCamelCase ( lowerCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE = ['input_ids', 'token_type_ids'] SCREAMING_SNAKE_CASE = FNetTokenizer def __init__(self , _lowerCamelCase=None , _lowerCamelCase=None , _lowerCamelCase=False , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase="<unk>" , _lowerCamelCase="[SEP]" , _lowerCamelCase="<pad>" , _lowerCamelCase="[CLS]" , _lowerCamelCase="[MASK]" , **_lowerCamelCase , ): """simple docstring""" UpperCAmelCase__ : Optional[Any] = ( AddedToken(_lowerCamelCase , lstrip=_lowerCamelCase , rstrip=_lowerCamelCase , normalized=_lowerCamelCase ) if isinstance(_lowerCamelCase , _lowerCamelCase ) else mask_token ) super().__init__( _lowerCamelCase , tokenizer_file=_lowerCamelCase , do_lower_case=_lowerCamelCase , remove_space=_lowerCamelCase , keep_accents=_lowerCamelCase , unk_token=_lowerCamelCase , sep_token=_lowerCamelCase , pad_token=_lowerCamelCase , cls_token=_lowerCamelCase , mask_token=_lowerCamelCase , **_lowerCamelCase , ) UpperCAmelCase__ : Optional[int] = do_lower_case UpperCAmelCase__ : List[str] = remove_space UpperCAmelCase__ : Optional[Any] = keep_accents UpperCAmelCase__ : List[str] = vocab_file UpperCAmelCase__ : Optional[int] = False if not self.vocab_file else True def _a (self , _lowerCamelCase , _lowerCamelCase = None ): """simple docstring""" UpperCAmelCase__ : Union[str, Any] = [self.sep_token_id] UpperCAmelCase__ : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _a (self , _lowerCamelCase , _lowerCamelCase = None ): """simple docstring""" UpperCAmelCase__ : List[str] = [self.sep_token_id] UpperCAmelCase__ : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _a (self , _lowerCamelCase , _lowerCamelCase = None ): """simple docstring""" if not os.path.isdir(_lowerCamelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase__ : List[str] = os.path.join( _lowerCamelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_lowerCamelCase ): copyfile(self.vocab_file , _lowerCamelCase ) return (out_vocab_file,)
166
0
from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging __A =logging.get_logger(__name__) # pylint: disable=invalid-name class _SCREAMING_SNAKE_CASE ( snake_case_ ): def __init__( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , ) -> List[Any]: super().__init__() if hasattr(scheduler.config , "steps_offset" ) and scheduler.config.steps_offset != 1: lowerCamelCase_ = ( f'The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`' f' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ' "to update the config accordingly as leaving `steps_offset` might led to incorrect results" " in future versions. If you have downloaded this checkpoint from the Hugging Face Hub," " it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`" " file" ) deprecate("steps_offset!=1" , "1.0.0" , lowercase , standard_warn=lowercase ) lowerCamelCase_ = dict(scheduler.config ) lowerCamelCase_ = 1 lowerCamelCase_ = FrozenDict(lowercase ) if hasattr(scheduler.config , "skip_prk_steps" ) and scheduler.config.skip_prk_steps is False: lowerCamelCase_ = ( f'The configuration file of this scheduler: {scheduler} has not set the configuration' " `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make" " sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to" " incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face" " Hub, it would be very nice if you could open a Pull request for the" " `scheduler/scheduler_config.json` file" ) deprecate("skip_prk_steps not set" , "1.0.0" , lowercase , standard_warn=lowercase ) lowerCamelCase_ = dict(scheduler.config ) lowerCamelCase_ = True lowerCamelCase_ = FrozenDict(lowercase ) if safety_checker is None: logger.warning( f'You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure' " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( segmentation_model=lowercase , segmentation_processor=lowercase , vae=lowercase , text_encoder=lowercase , tokenizer=lowercase , unet=lowercase , scheduler=lowercase , safety_checker=lowercase , feature_extractor=lowercase , ) def SCREAMING_SNAKE_CASE_( self , lowercase = "auto" ) -> Tuple: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory lowerCamelCase_ = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> List[Any]: self.enable_attention_slicing(lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> str: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) lowerCamelCase_ = torch.device("cuda" ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(lowercase , lowercase ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def SCREAMING_SNAKE_CASE_( self ) -> Union[str, Any]: if self.device != torch.device("meta" ) or not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowercase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self , lowercase , lowercase , lowercase , lowercase = 512 , lowercase = 512 , lowercase = 50 , lowercase = 7.5 , lowercase = None , lowercase = 1 , lowercase = 0.0 , lowercase = None , lowercase = None , lowercase = "pil" , lowercase = True , lowercase = None , lowercase = 1 , **lowercase , ) -> int: lowerCamelCase_ = self.segmentation_processor( text=[text] , images=[image] , padding="max_length" , return_tensors="pt" ).to(self.device ) lowerCamelCase_ = self.segmentation_model(**lowercase ) lowerCamelCase_ = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() lowerCamelCase_ = self.numpy_to_pil(lowercase )[0].resize(image.size ) # Run inpainting pipeline with the generated mask lowerCamelCase_ = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=lowercase , image=lowercase , mask_image=lowercase , height=lowercase , width=lowercase , num_inference_steps=lowercase , guidance_scale=lowercase , negative_prompt=lowercase , num_images_per_prompt=lowercase , eta=lowercase , generator=lowercase , latents=lowercase , output_type=lowercase , return_dict=lowercase , callback=lowercase , callback_steps=lowercase , )
19
def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ ): lowerCamelCase_ = [0 for i in range(r + 1 )] # nc0 = 1 lowerCamelCase_ = 1 for i in range(1 , n + 1 ): # to compute current row from previous row. lowerCamelCase_ = min(lowerCamelCase__ , lowerCamelCase__ ) while j > 0: c[j] += c[j - 1] j -= 1 return c[r] print(binomial_coefficient(n=1_0, r=5))
19
1
'''simple docstring''' def a__ ( _SCREAMING_SNAKE_CASE : Optional[int] ) -> Any: return [ { 0: [1, 2], 1: [0, 2], 2: [0, 1, 3, 5], 3: [2, 4], 4: [3], 5: [2, 6, 8], 6: [5, 7], 7: [6, 8], 8: [5, 7], }, { 0: [6], 1: [9], 2: [4, 5], 3: [4], 4: [2, 3], 5: [2], 6: [0, 7], 7: [6], 8: [], 9: [1], }, { 0: [4], 1: [6], 2: [], 3: [5, 6, 7], 4: [0, 6], 5: [3, 8, 9], 6: [1, 3, 4, 7], 7: [3, 6, 8, 9], 8: [5, 7], 9: [5, 7], }, { 0: [1, 3], 1: [0, 2, 4], 2: [1, 3, 4], 3: [0, 2, 4], 4: [1, 2, 3], }, ][index] def a__ ( _SCREAMING_SNAKE_CASE : dict[int, list[int]] ) -> list[tuple[int, int]]: UpperCAmelCase_ : Any = 0 UpperCAmelCase_ : int = len(_SCREAMING_SNAKE_CASE ) # No of vertices in graph UpperCAmelCase_ : int = [0] * n UpperCAmelCase_ : Union[str, Any] = [False] * n def dfs(_SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : Dict , _SCREAMING_SNAKE_CASE : List[Any] , _SCREAMING_SNAKE_CASE : Union[str, Any] ): UpperCAmelCase_ : Any = True UpperCAmelCase_ : str = id_ id_ += 1 for to in graph[at]: if to == parent: pass elif not visited[to]: dfs(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , id_ ) UpperCAmelCase_ : Dict = min(low[at] , low[to] ) if id_ <= low[to]: bridges.append((at, to) if at < to else (to, at) ) else: # This edge is a back edge and cannot be a bridge UpperCAmelCase_ : int = min(low[at] , low[to] ) UpperCAmelCase_ : list[tuple[int, int]] = [] for i in range(_SCREAMING_SNAKE_CASE ): if not visited[i]: dfs(_SCREAMING_SNAKE_CASE , -1 , _SCREAMING_SNAKE_CASE , id_ ) return bridges if __name__ == "__main__": import doctest doctest.testmod()
368
'''simple docstring''' from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class _snake_case : def __init__( self ,_snake_case ,_snake_case=12 ,_snake_case=7 ,_snake_case=True ,_snake_case=True ,_snake_case=True ,_snake_case=99 ,_snake_case=32 ,_snake_case=32 ,_snake_case=2 ,_snake_case=4 ,_snake_case=37 ,_snake_case=0.1 ,_snake_case=0.1 ,_snake_case=5_12 ,_snake_case=0.02 ,_snake_case=0 ,_snake_case=None ,): UpperCAmelCase_ : Any = parent UpperCAmelCase_ : int = batch_size UpperCAmelCase_ : Optional[int] = seq_length UpperCAmelCase_ : Union[str, Any] = is_training UpperCAmelCase_ : str = use_input_mask UpperCAmelCase_ : List[Any] = use_labels UpperCAmelCase_ : Union[str, Any] = vocab_size UpperCAmelCase_ : List[Any] = hidden_size UpperCAmelCase_ : Any = projection_dim UpperCAmelCase_ : Optional[Any] = num_hidden_layers UpperCAmelCase_ : Union[str, Any] = num_attention_heads UpperCAmelCase_ : Optional[Any] = intermediate_size UpperCAmelCase_ : Any = dropout UpperCAmelCase_ : Dict = attention_dropout UpperCAmelCase_ : Optional[Any] = max_position_embeddings UpperCAmelCase_ : List[Any] = initializer_range UpperCAmelCase_ : Optional[int] = scope UpperCAmelCase_ : List[str] = bos_token_id def UpperCamelCase__ ( self ): UpperCAmelCase_ : Dict = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase_ : List[Any] = None if self.use_input_mask: UpperCAmelCase_ : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: UpperCAmelCase_ : Any = input_mask.numpy() UpperCAmelCase_ , UpperCAmelCase_ : str = input_mask.shape UpperCAmelCase_ : str = np.random.randint(1 ,seq_length - 1 ,size=(batch_size,) ) for batch_idx, start_index in enumerate(_snake_case ): UpperCAmelCase_ : Optional[int] = 1 UpperCAmelCase_ : Union[str, Any] = 0 UpperCAmelCase_ : int = self.get_config() return config, input_ids, tf.convert_to_tensor(_snake_case ) def UpperCamelCase__ ( self ): return BlipTextConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,projection_dim=self.projection_dim ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,dropout=self.dropout ,attention_dropout=self.attention_dropout ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,bos_token_id=self.bos_token_id ,) def UpperCamelCase__ ( self ,_snake_case ,_snake_case ,_snake_case ): UpperCAmelCase_ : Union[str, Any] = TFBlipTextModel(config=_snake_case ) UpperCAmelCase_ : Optional[int] = model(_snake_case ,attention_mask=_snake_case ,training=_snake_case ) UpperCAmelCase_ : Dict = model(_snake_case ,training=_snake_case ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape ,(self.batch_size, self.hidden_size) ) def UpperCamelCase__ ( self ): UpperCAmelCase_ : Optional[int] = self.prepare_config_and_inputs() UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : str = config_and_inputs UpperCAmelCase_ : str = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class _snake_case (__SCREAMING_SNAKE_CASE , unittest.TestCase): __A : Tuple =(TFBlipTextModel,) if is_tf_available() else () __A : List[Any] =False __A : List[Any] =False __A : Any =False def UpperCamelCase__ ( self ): UpperCAmelCase_ : Any = BlipTextModelTester(self ) UpperCAmelCase_ : List[Any] = ConfigTester(self ,config_class=_snake_case ,hidden_size=37 ) def UpperCamelCase__ ( self ): self.config_tester.run_common_tests() def UpperCamelCase__ ( self ): UpperCAmelCase_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_snake_case ) def UpperCamelCase__ ( self ): pass def UpperCamelCase__ ( self ): pass @unittest.skip(reason="Blip does not use inputs_embeds" ) def UpperCamelCase__ ( self ): pass @unittest.skip(reason="BlipTextModel has no base class and is not available in MODEL_MAPPING" ) def UpperCamelCase__ ( self ): pass @unittest.skip(reason="BlipTextModel has no base class and is not available in MODEL_MAPPING" ) def UpperCamelCase__ ( self ): pass @slow def UpperCamelCase__ ( self ): for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ : int = TFBlipTextModel.from_pretrained(_snake_case ) self.assertIsNotNone(_snake_case ) def UpperCamelCase__ ( self ,_snake_case=True ): super().test_pt_tf_model_equivalence(allow_missing_keys=_snake_case )
67
0
from ....configuration_utils import PretrainedConfig from ....utils import logging UpperCAmelCase : Any = logging.get_logger(__name__) # TODO: upload to AWS UpperCAmelCase : Dict = { """yjernite/retribert-base-uncased""": ( """https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/config.json""" ), } class __lowerCAmelCase ( UpperCamelCase__): _lowercase : Optional[Any] = """retribert""" def __init__( self , lowerCAmelCase__=3_0_5_2_2 , lowerCAmelCase__=7_6_8 , lowerCAmelCase__=8 , lowerCAmelCase__=1_2 , lowerCAmelCase__=3_0_7_2 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=5_1_2 , lowerCAmelCase__=2 , lowerCAmelCase__=0.02 , lowerCAmelCase__=1E-12 , lowerCAmelCase__=True , lowerCAmelCase__=1_2_8 , lowerCAmelCase__=0 , **lowerCAmelCase__ , ) -> str: '''simple docstring''' super().__init__(pad_token_id=lowerCAmelCase__ , **lowerCAmelCase__ ) a__ : Tuple =vocab_size a__ : Any =hidden_size a__ : Optional[Any] =num_hidden_layers a__ : Optional[int] =num_attention_heads a__ : List[Any] =hidden_act a__ : Dict =intermediate_size a__ : Optional[int] =hidden_dropout_prob a__ : Dict =attention_probs_dropout_prob a__ : Optional[int] =max_position_embeddings a__ : List[str] =type_vocab_size a__ : Union[str, Any] =initializer_range a__ : Optional[Any] =layer_norm_eps a__ : int =share_encoders a__ : List[str] =projection_dim
95
from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxSeqaSeqConfigWithPast from ...utils import logging a : Tuple = logging.get_logger(__name__) a : Dict = { "t5-small": "https://huggingface.co/t5-small/resolve/main/config.json", "t5-base": "https://huggingface.co/t5-base/resolve/main/config.json", "t5-large": "https://huggingface.co/t5-large/resolve/main/config.json", "t5-3b": "https://huggingface.co/t5-3b/resolve/main/config.json", "t5-11b": "https://huggingface.co/t5-11b/resolve/main/config.json", } class a ( lowercase__ ): """simple docstring""" a : int = 't5' a : Dict = ['past_key_values'] a : Tuple = {'hidden_size': 'd_model', 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers'} def __init__( self : str , __lowercase : Optional[int]=32128 , __lowercase : Optional[int]=512 , __lowercase : int=64 , __lowercase : Any=2048 , __lowercase : Tuple=6 , __lowercase : Tuple=None , __lowercase : int=8 , __lowercase : List[Any]=32 , __lowercase : Dict=128 , __lowercase : Optional[int]=0.1 , __lowercase : int=1e-6 , __lowercase : List[str]=1.0 , __lowercase : List[str]="relu" , __lowercase : Dict=True , __lowercase : Optional[Any]=True , __lowercase : Tuple=0 , __lowercase : List[str]=1 , **__lowercase : Any , ) -> str: __UpperCAmelCase : Tuple = vocab_size __UpperCAmelCase : Optional[int] = d_model __UpperCAmelCase : Union[str, Any] = d_kv __UpperCAmelCase : Union[str, Any] = d_ff __UpperCAmelCase : int = num_layers __UpperCAmelCase : Any = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry __UpperCAmelCase : Dict = num_heads __UpperCAmelCase : List[Any] = relative_attention_num_buckets __UpperCAmelCase : List[str] = relative_attention_max_distance __UpperCAmelCase : Union[str, Any] = dropout_rate __UpperCAmelCase : List[str] = layer_norm_epsilon __UpperCAmelCase : str = initializer_factor __UpperCAmelCase : Dict = feed_forward_proj __UpperCAmelCase : Optional[int] = use_cache __UpperCAmelCase : List[Any] = self.feed_forward_proj.split("""-""" ) __UpperCAmelCase : Tuple = act_info[-1] __UpperCAmelCase : int = act_info[0] == """gated""" if len(__lowercase ) > 1 and act_info[0] != "gated" or len(__lowercase ) > 2: raise ValueError( f"""`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.""" """Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. """ """'gated-gelu' or 'relu'""" ) # for backwards compatibility if feed_forward_proj == "gated-gelu": __UpperCAmelCase : Dict = """gelu_new""" super().__init__( pad_token_id=__lowercase , eos_token_id=__lowercase , is_encoder_decoder=__lowercase , **__lowercase , ) class a ( lowercase__ ): """simple docstring""" @property def UpperCAmelCase ( self : str ) -> Mapping[str, Mapping[int, str]]: __UpperCAmelCase : Union[str, Any] = { """input_ids""": {0: """batch""", 1: """encoder_sequence"""}, """attention_mask""": {0: """batch""", 1: """encoder_sequence"""}, } if self.use_past: __UpperCAmelCase : List[Any] = """past_encoder_sequence + sequence""" __UpperCAmelCase : Optional[int] = {0: """batch"""} __UpperCAmelCase : Union[str, Any] = {0: """batch""", 1: """past_decoder_sequence + sequence"""} else: __UpperCAmelCase : str = {0: """batch""", 1: """decoder_sequence"""} __UpperCAmelCase : str = {0: """batch""", 1: """decoder_sequence"""} if self.use_past: self.fill_with_past_key_values_(__lowercase , direction="""inputs""" ) return common_inputs @property def UpperCAmelCase ( self : int ) -> int: return 13
114
0
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { 'facebook/convnextv2-tiny-1k-224': 'https://huggingface.co/facebook/convnextv2-tiny-1k-224/resolve/main/config.json', } class a_ ( snake_case_ , snake_case_ ): '''simple docstring''' UpperCAmelCase_ = 'convnextv2' def __init__( self : Optional[int] , lowercase__ : Dict=3 , lowercase__ : Optional[Any]=4 , lowercase__ : Tuple=4 , lowercase__ : Optional[int]=None , lowercase__ : List[Any]=None , lowercase__ : int="gelu" , lowercase__ : List[Any]=0.02 , lowercase__ : int=1e-12 , lowercase__ : List[str]=0.0 , lowercase__ : List[Any]=224 , lowercase__ : Any=None , lowercase__ : List[str]=None , **lowercase__ : Optional[int] , ): '''simple docstring''' super().__init__(**lowercase__) lowerCAmelCase__ = num_channels lowerCAmelCase__ = patch_size lowerCAmelCase__ = num_stages lowerCAmelCase__ = [96, 192, 384, 768] if hidden_sizes is None else hidden_sizes lowerCAmelCase__ = [3, 3, 9, 3] if depths is None else depths lowerCAmelCase__ = hidden_act lowerCAmelCase__ = initializer_range lowerCAmelCase__ = layer_norm_eps lowerCAmelCase__ = drop_path_rate lowerCAmelCase__ = image_size lowerCAmelCase__ = ['stem'] + [F"""stage{idx}""" for idx in range(1 , len(self.depths) + 1)] lowerCAmelCase__ , lowerCAmelCase__ = get_aligned_output_features_output_indices( out_features=lowercase__ , out_indices=lowercase__ , stage_names=self.stage_names)
353
def __lowerCamelCase ( lowerCAmelCase__ ): lowerCAmelCase__ = len(lowerCAmelCase__ ) for i in range(lowerCAmelCase__ ): for j in range(i + 1 , lowerCAmelCase__ ): if numbers[j] < numbers[i]: lowerCAmelCase__ , lowerCAmelCase__ = numbers[j], numbers[i] return numbers if __name__ == "__main__": lowerCAmelCase__ = input('Enter numbers separated by a comma:\n').strip() lowerCAmelCase__ = [int(item) for item in user_input.split(',')] print(exchange_sort(unsorted))
119
0
from ..utils import DummyObject, requires_backends class a ( metaclass=lowercase__ ): """simple docstring""" a : Tuple = ['note_seq'] def __init__( self : int , *__lowercase : Optional[int] , **__lowercase : Optional[int] ) -> str: requires_backends(self , ["""note_seq"""] ) @classmethod def UpperCAmelCase ( cls : str , *__lowercase : str , **__lowercase : str ) -> Any: requires_backends(cls , ["""note_seq"""] ) @classmethod def UpperCAmelCase ( cls : str , *__lowercase : List[Any] , **__lowercase : Tuple ) -> Tuple: requires_backends(cls , ["""note_seq"""] )
114
import random import unittest import numpy as np from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionImgaImgPipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class a ( lowercase__ , unittest.TestCase ): """simple docstring""" a : int = 'hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline' def UpperCAmelCase ( self : List[str] , __lowercase : Optional[Any]=0 ) -> Any: __UpperCAmelCase : Any = floats_tensor((1, 3, 128, 128) , rng=random.Random(__lowercase ) ) __UpperCAmelCase : int = np.random.RandomState(__lowercase ) __UpperCAmelCase : Optional[Any] = { """prompt""": """A painting of a squirrel eating a burger""", """image""": image, """generator""": generator, """num_inference_steps""": 3, """strength""": 0.75, """guidance_scale""": 7.5, """output_type""": """numpy""", } return inputs def UpperCAmelCase ( self : Optional[int] ) -> Optional[Any]: __UpperCAmelCase : List[Any] = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) pipe.set_progress_bar_config(disable=__lowercase ) __UpperCAmelCase : int = self.get_dummy_inputs() __UpperCAmelCase : Optional[Any] = pipe(**__lowercase ).images __UpperCAmelCase : Union[str, Any] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 128, 128, 3) __UpperCAmelCase : List[str] = np.array([0.69_643, 0.58_484, 0.50_314, 0.58_760, 0.55_368, 0.59_643, 0.51_529, 0.41_217, 0.49_087] ) assert np.abs(image_slice - expected_slice ).max() < 1e-1 def UpperCAmelCase ( self : Optional[Any] ) -> List[str]: __UpperCAmelCase : Dict = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __UpperCAmelCase : int = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=__lowercase ) pipe.set_progress_bar_config(disable=__lowercase ) __UpperCAmelCase : Any = self.get_dummy_inputs() __UpperCAmelCase : Tuple = pipe(**__lowercase ).images __UpperCAmelCase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __UpperCAmelCase : str = np.array([0.61_737, 0.54_642, 0.53_183, 0.54_465, 0.52_742, 0.60_525, 0.49_969, 0.40_655, 0.48_154] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def UpperCAmelCase ( self : str ) -> Tuple: __UpperCAmelCase : Tuple = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __UpperCAmelCase : int = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__lowercase ) # warmup pass to apply optimizations __UpperCAmelCase : Optional[int] = pipe(**self.get_dummy_inputs() ) __UpperCAmelCase : Tuple = self.get_dummy_inputs() __UpperCAmelCase : Any = pipe(**__lowercase ).images __UpperCAmelCase : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __UpperCAmelCase : Optional[int] = np.array([0.52_761, 0.59_977, 0.49_033, 0.49_619, 0.54_282, 0.50_311, 0.47_600, 0.40_918, 0.45_203] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def UpperCAmelCase ( self : Optional[Any] ) -> str: __UpperCAmelCase : int = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __UpperCAmelCase : Union[str, Any] = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__lowercase ) __UpperCAmelCase : List[str] = self.get_dummy_inputs() __UpperCAmelCase : int = pipe(**__lowercase ).images __UpperCAmelCase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __UpperCAmelCase : Tuple = np.array([0.52_911, 0.60_004, 0.49_229, 0.49_805, 0.54_502, 0.50_680, 0.47_777, 0.41_028, 0.45_304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def UpperCAmelCase ( self : int ) -> Any: __UpperCAmelCase : Any = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __UpperCAmelCase : List[Any] = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__lowercase ) __UpperCAmelCase : Union[str, Any] = self.get_dummy_inputs() __UpperCAmelCase : int = pipe(**__lowercase ).images __UpperCAmelCase : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __UpperCAmelCase : List[str] = np.array([0.52_911, 0.60_004, 0.49_229, 0.49_805, 0.54_502, 0.50_680, 0.47_777, 0.41_028, 0.45_304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def UpperCAmelCase ( self : Tuple ) -> str: __UpperCAmelCase : Dict = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __UpperCAmelCase : str = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__lowercase ) __UpperCAmelCase : Optional[Any] = self.get_dummy_inputs() __UpperCAmelCase : int = pipe(**__lowercase ).images __UpperCAmelCase : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __UpperCAmelCase : Union[str, Any] = np.array([0.65_331, 0.58_277, 0.48_204, 0.56_059, 0.53_665, 0.56_235, 0.50_969, 0.40_009, 0.46_552] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 @nightly @require_onnxruntime @require_torch_gpu class a ( unittest.TestCase ): """simple docstring""" @property def UpperCAmelCase ( self : Dict ) -> List[Any]: return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCAmelCase ( self : Tuple ) -> Tuple: __UpperCAmelCase : Optional[int] = ort.SessionOptions() __UpperCAmelCase : List[Any] = False return options def UpperCAmelCase ( self : List[str] ) -> Tuple: __UpperCAmelCase : List[Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/img2img/sketch-mountains-input.jpg""" ) __UpperCAmelCase : Dict = init_image.resize((768, 512) ) # using the PNDM scheduler by default __UpperCAmelCase : Any = OnnxStableDiffusionImgaImgPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""onnx""" , safety_checker=__lowercase , feature_extractor=__lowercase , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=__lowercase ) __UpperCAmelCase : Dict = """A fantasy landscape, trending on artstation""" __UpperCAmelCase : str = np.random.RandomState(0 ) __UpperCAmelCase : Optional[Any] = pipe( prompt=__lowercase , image=__lowercase , strength=0.75 , guidance_scale=7.5 , num_inference_steps=10 , generator=__lowercase , output_type="""np""" , ) __UpperCAmelCase : str = output.images __UpperCAmelCase : int = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 768, 3) __UpperCAmelCase : Union[str, Any] = np.array([0.4_909, 0.5_059, 0.5_372, 0.4_623, 0.4_876, 0.5_049, 0.4_820, 0.4_956, 0.5_019] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 def UpperCAmelCase ( self : Optional[Any] ) -> str: __UpperCAmelCase : Tuple = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/img2img/sketch-mountains-input.jpg""" ) __UpperCAmelCase : int = init_image.resize((768, 512) ) __UpperCAmelCase : Tuple = LMSDiscreteScheduler.from_pretrained( """runwayml/stable-diffusion-v1-5""" , subfolder="""scheduler""" , revision="""onnx""" ) __UpperCAmelCase : Tuple = OnnxStableDiffusionImgaImgPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , revision="""onnx""" , scheduler=__lowercase , safety_checker=__lowercase , feature_extractor=__lowercase , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=__lowercase ) __UpperCAmelCase : Dict = """A fantasy landscape, trending on artstation""" __UpperCAmelCase : int = np.random.RandomState(0 ) __UpperCAmelCase : Optional[int] = pipe( prompt=__lowercase , image=__lowercase , strength=0.75 , guidance_scale=7.5 , num_inference_steps=20 , generator=__lowercase , output_type="""np""" , ) __UpperCAmelCase : Union[str, Any] = output.images __UpperCAmelCase : Union[str, Any] = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 768, 3) __UpperCAmelCase : str = np.array([0.8_043, 0.926, 0.9_581, 0.8_119, 0.8_954, 0.913, 0.7_209, 0.7_463, 0.7_431] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2
114
1
from maths.prime_factors import prime_factors def lowercase__ ( __snake_case : int ): '''simple docstring''' if not isinstance(__snake_case , __snake_case ): UpperCAmelCase_ : Optional[Any] = F"Input value of [number={number}] must be an integer" raise TypeError(__snake_case ) if number < 1: raise ValueError('Input must be a positive integer' ) return -1 if len(prime_factors(__snake_case ) ) % 2 else 1 if __name__ == "__main__": import doctest doctest.testmod()
145
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...onnx.utils import compute_effective_axis_dimension from ...utils import logging if TYPE_CHECKING: from ...processing_utils import ProcessorMixin from ...utils import TensorType __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { 'microsoft/layoutlmv3-base': 'https://huggingface.co/microsoft/layoutlmv3-base/resolve/main/config.json', } class lowerCamelCase (_snake_case ): '''simple docstring''' _snake_case : Dict = '''layoutlmv3''' def __init__( self , _UpperCamelCase=5_0_2_6_5 , _UpperCamelCase=7_6_8 , _UpperCamelCase=1_2 , _UpperCamelCase=1_2 , _UpperCamelCase=3_0_7_2 , _UpperCamelCase="gelu" , _UpperCamelCase=0.1 , _UpperCamelCase=0.1 , _UpperCamelCase=5_1_2 , _UpperCamelCase=2 , _UpperCamelCase=0.02 , _UpperCamelCase=1E-5 , _UpperCamelCase=1 , _UpperCamelCase=0 , _UpperCamelCase=2 , _UpperCamelCase=1_0_2_4 , _UpperCamelCase=1_2_8 , _UpperCamelCase=1_2_8 , _UpperCamelCase=True , _UpperCamelCase=3_2 , _UpperCamelCase=1_2_8 , _UpperCamelCase=6_4 , _UpperCamelCase=2_5_6 , _UpperCamelCase=True , _UpperCamelCase=True , _UpperCamelCase=True , _UpperCamelCase=2_2_4 , _UpperCamelCase=3 , _UpperCamelCase=1_6 , _UpperCamelCase=None , **_UpperCamelCase , ) -> Optional[Any]: super().__init__( vocab_size=_UpperCamelCase , hidden_size=_UpperCamelCase , num_hidden_layers=_UpperCamelCase , num_attention_heads=_UpperCamelCase , intermediate_size=_UpperCamelCase , hidden_act=_UpperCamelCase , hidden_dropout_prob=_UpperCamelCase , attention_probs_dropout_prob=_UpperCamelCase , max_position_embeddings=_UpperCamelCase , type_vocab_size=_UpperCamelCase , initializer_range=_UpperCamelCase , layer_norm_eps=_UpperCamelCase , pad_token_id=_UpperCamelCase , bos_token_id=_UpperCamelCase , eos_token_id=_UpperCamelCase , **_UpperCamelCase , ) UpperCAmelCase_ : str = max_ad_position_embeddings UpperCAmelCase_ : Union[str, Any] = coordinate_size UpperCAmelCase_ : Union[str, Any] = shape_size UpperCAmelCase_ : str = has_relative_attention_bias UpperCAmelCase_ : Tuple = rel_pos_bins UpperCAmelCase_ : Dict = max_rel_pos UpperCAmelCase_ : Any = has_spatial_attention_bias UpperCAmelCase_ : Optional[Any] = rel_ad_pos_bins UpperCAmelCase_ : List[str] = max_rel_ad_pos UpperCAmelCase_ : List[str] = text_embed UpperCAmelCase_ : Dict = visual_embed UpperCAmelCase_ : Optional[int] = input_size UpperCAmelCase_ : List[Any] = num_channels UpperCAmelCase_ : List[Any] = patch_size UpperCAmelCase_ : Union[str, Any] = classifier_dropout class lowerCamelCase (_snake_case ): '''simple docstring''' _snake_case : Tuple = version.parse('''1.12''' ) @property def __UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: # The order of inputs is different for question answering and sequence classification if self.task in ["question-answering", "sequence-classification"]: return OrderedDict( [ ('input_ids', {0: 'batch', 1: 'sequence'}), ('attention_mask', {0: 'batch', 1: 'sequence'}), ('bbox', {0: 'batch', 1: 'sequence'}), ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) else: return OrderedDict( [ ('input_ids', {0: 'batch', 1: 'sequence'}), ('bbox', {0: 'batch', 1: 'sequence'}), ('attention_mask', {0: 'batch', 1: 'sequence'}), ('pixel_values', {0: 'batch', 1: 'num_channels'}), ] ) @property def __UpperCAmelCase ( self ) -> float: return 1E-5 @property def __UpperCAmelCase ( self ) -> int: return 1_2 def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase = -1 , _UpperCamelCase = -1 , _UpperCamelCase = False , _UpperCamelCase = None , _UpperCamelCase = 3 , _UpperCamelCase = 4_0 , _UpperCamelCase = 4_0 , ) -> Mapping[str, Any]: setattr(processor.image_processor , 'apply_ocr' , _UpperCamelCase ) # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX UpperCAmelCase_ : Optional[int] = compute_effective_axis_dimension( _UpperCamelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX UpperCAmelCase_ : Any = processor.tokenizer.num_special_tokens_to_add(_UpperCamelCase ) UpperCAmelCase_ : Any = compute_effective_axis_dimension( _UpperCamelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=_UpperCamelCase ) # Generate dummy inputs according to compute batch and sequence UpperCAmelCase_ : Tuple = [[' '.join([processor.tokenizer.unk_token] ) * seq_length]] * batch_size # Generate dummy bounding boxes UpperCAmelCase_ : List[str] = [[[4_8, 8_4, 7_3, 1_2_8]]] * batch_size # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX # batch_size = compute_effective_axis_dimension(batch_size, fixed_dimension=OnnxConfig.default_fixed_batch) UpperCAmelCase_ : str = self._generate_dummy_images(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) UpperCAmelCase_ : List[str] = dict( processor( _UpperCamelCase , text=_UpperCamelCase , boxes=_UpperCamelCase , return_tensors=_UpperCamelCase , ) ) return inputs
145
1
"""simple docstring""" from __future__ import annotations def _snake_case ( lowerCamelCase__ : int | float | str , lowerCamelCase__ : int | float | str ) -> list[str]: if nth_term == "": return [""] lowerCamelCase_ : Any =int(lowerCamelCase__ ) lowerCamelCase_ : int =int(lowerCamelCase__ ) lowerCamelCase_ : list[str] =[] for temp in range(int(lowerCamelCase__ ) ): series.append(F"""1 / {pow(temp + 1 , int(lowerCamelCase__ ) )}""" if series else "1" ) return series if __name__ == "__main__": import doctest doctest.testmod() A__ : Optional[int] = int(input('Enter the last number (nth term) of the P-Series')) A__ : List[Any] = int(input('Enter the power for P-Series')) print('Formula of P-Series => 1+1/2^p+1/3^p ..... 1/n^p') print(p_series(nth_term, power))
144
"""simple docstring""" A__ : Optional[int] = { 'Pillow': 'Pillow<10.0.0', 'accelerate': 'accelerate>=0.20.3', 'av': 'av==9.2.0', 'beautifulsoup4': 'beautifulsoup4', 'black': 'black~=23.1', 'codecarbon': 'codecarbon==1.2.0', 'cookiecutter': 'cookiecutter==1.7.3', 'dataclasses': 'dataclasses', 'datasets': 'datasets!=2.5.0', 'decord': 'decord==0.6.0', 'deepspeed': 'deepspeed>=0.9.3', 'diffusers': 'diffusers', 'dill': 'dill<0.3.5', 'evaluate': 'evaluate>=0.2.0', 'fairscale': 'fairscale>0.3', 'faiss-cpu': 'faiss-cpu', 'fastapi': 'fastapi', 'filelock': 'filelock', 'flax': 'flax>=0.4.1,<=0.7.0', 'ftfy': 'ftfy', 'fugashi': 'fugashi>=1.0', 'GitPython': 'GitPython<3.1.19', 'hf-doc-builder': 'hf-doc-builder>=0.3.0', 'huggingface-hub': 'huggingface-hub>=0.14.1,<1.0', 'importlib_metadata': 'importlib_metadata', 'ipadic': 'ipadic>=1.0.0,<2.0', 'isort': 'isort>=5.5.4', 'jax': 'jax>=0.2.8,!=0.3.2,<=0.4.13', 'jaxlib': 'jaxlib>=0.1.65,<=0.4.13', 'jieba': 'jieba', 'kenlm': 'kenlm', 'keras-nlp': 'keras-nlp>=0.3.1', 'librosa': 'librosa', 'nltk': 'nltk', 'natten': 'natten>=0.14.6', 'numpy': 'numpy>=1.17', 'onnxconverter-common': 'onnxconverter-common', 'onnxruntime-tools': 'onnxruntime-tools>=1.4.2', 'onnxruntime': 'onnxruntime>=1.4.0', 'opencv-python': 'opencv-python', 'optuna': 'optuna', 'optax': 'optax>=0.0.8,<=0.1.4', 'packaging': 'packaging>=20.0', 'parameterized': 'parameterized', 'phonemizer': 'phonemizer', 'protobuf': 'protobuf', 'psutil': 'psutil', 'pyyaml': 'pyyaml>=5.1', 'pydantic': 'pydantic<2', 'pytest': 'pytest>=7.2.0', 'pytest-timeout': 'pytest-timeout', 'pytest-xdist': 'pytest-xdist', 'python': 'python>=3.8.0', 'ray[tune]': 'ray[tune]', 'regex': 'regex!=2019.12.17', 'requests': 'requests', 'rhoknp': 'rhoknp>=1.1.0,<1.3.1', 'rjieba': 'rjieba', 'rouge-score': 'rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1', 'ruff': 'ruff>=0.0.241,<=0.0.259', 'sacrebleu': 'sacrebleu>=1.4.12,<2.0.0', 'sacremoses': 'sacremoses', 'safetensors': 'safetensors>=0.3.1', 'sagemaker': 'sagemaker>=2.31.0', 'scikit-learn': 'scikit-learn', 'sentencepiece': 'sentencepiece>=0.1.91,!=0.1.92', 'sigopt': 'sigopt', 'starlette': 'starlette', 'sudachipy': 'sudachipy>=0.6.6', 'sudachidict_core': 'sudachidict_core>=20220729', 'tensorflow-cpu': 'tensorflow-cpu>=2.6,<2.14', 'tensorflow': 'tensorflow>=2.6,<2.14', 'tensorflow-text': 'tensorflow-text<2.14', 'tf2onnx': 'tf2onnx', 'timeout-decorator': 'timeout-decorator', 'timm': 'timm', 'tokenizers': 'tokenizers>=0.11.1,!=0.11.3,<0.14', 'torch': 'torch>=1.9,!=1.12.0', 'torchaudio': 'torchaudio', 'torchvision': 'torchvision', 'pyctcdecode': 'pyctcdecode>=0.4.0', 'tqdm': 'tqdm>=4.27', 'unidic': 'unidic>=1.0.2', 'unidic_lite': 'unidic_lite>=1.0.7', 'urllib3': 'urllib3<2.0.0', 'uvicorn': 'uvicorn', }
144
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _lowerCAmelCase : int = { '''configuration_groupvit''': [ '''GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GroupViTConfig''', '''GroupViTOnnxConfig''', '''GroupViTTextConfig''', '''GroupViTVisionConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : Tuple = [ '''GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GroupViTModel''', '''GroupViTPreTrainedModel''', '''GroupViTTextModel''', '''GroupViTVisionModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : Optional[int] = [ '''TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFGroupViTModel''', '''TFGroupViTPreTrainedModel''', '''TFGroupViTTextModel''', '''TFGroupViTVisionModel''', ] if TYPE_CHECKING: from .configuration_groupvit import ( GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GroupViTConfig, GroupViTOnnxConfig, GroupViTTextConfig, GroupViTVisionConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_groupvit import ( GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, GroupViTModel, GroupViTPreTrainedModel, GroupViTTextModel, GroupViTVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_groupvit import ( TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFGroupViTModel, TFGroupViTPreTrainedModel, TFGroupViTTextModel, TFGroupViTVisionModel, ) else: import sys _lowerCAmelCase : List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
350
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _SCREAMING_SNAKE_CASE ( __a ,__a ,__a ,unittest.TestCase ): __SCREAMING_SNAKE_CASE :Union[str, Any] = StableDiffusionInpaintPipeline __SCREAMING_SNAKE_CASE :Optional[int] = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS __SCREAMING_SNAKE_CASE :Tuple = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS __SCREAMING_SNAKE_CASE :str = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess __SCREAMING_SNAKE_CASE :Optional[Any] = frozenset([] ) def snake_case__ ( self : Union[str, Any] ): torch.manual_seed(0 ) __magic_name__ = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=9 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=a__ , ) __magic_name__ = PNDMScheduler(skip_prk_steps=a__ ) torch.manual_seed(0 ) __magic_name__ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) __magic_name__ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='''gelu''' , projection_dim=512 , ) __magic_name__ = CLIPTextModel(a__ ) __magic_name__ = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) __magic_name__ = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def snake_case__ ( self : Any , a__ : Optional[int] , a__ : List[Any]=0 ): # TODO: use tensor inputs instead of PIL, this is here just to leave the old expected_slices untouched __magic_name__ = floats_tensor((1, 3, 32, 32) , rng=random.Random(a__ ) ).to(a__ ) __magic_name__ = image.cpu().permute(0 , 2 , 3 , 1 )[0] __magic_name__ = Image.fromarray(np.uinta(a__ ) ).convert('''RGB''' ).resize((64, 64) ) __magic_name__ = Image.fromarray(np.uinta(image + 4 ) ).convert('''RGB''' ).resize((64, 64) ) if str(a__ ).startswith('''mps''' ): __magic_name__ = torch.manual_seed(a__ ) else: __magic_name__ = torch.Generator(device=a__ ).manual_seed(a__ ) __magic_name__ = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': init_image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def snake_case__ ( self : Union[str, Any] ): __magic_name__ = '''cpu''' # ensure determinism for the device-dependent torch.Generator __magic_name__ = self.get_dummy_components() __magic_name__ = StableDiffusionInpaintPipeline(**a__ ) __magic_name__ = sd_pipe.to(a__ ) sd_pipe.set_progress_bar_config(disable=a__ ) __magic_name__ = self.get_dummy_inputs(a__ ) __magic_name__ = sd_pipe(**a__ ).images __magic_name__ = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) __magic_name__ = np.array([0.4_727, 0.5_735, 0.3_941, 0.5_446, 0.5_926, 0.4_394, 0.5_062, 0.4_654, 0.4_476] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def snake_case__ ( self : List[Any] ): super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): def snake_case__ ( self : int ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case__ ( self : Tuple ): __magic_name__ = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''' ) __magic_name__ = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''' ) __magic_name__ = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint''' '''/yellow_cat_sitting_on_a_park_bench.npy''' ) __magic_name__ = '''stabilityai/stable-diffusion-2-inpainting''' __magic_name__ = StableDiffusionInpaintPipeline.from_pretrained(a__ , safety_checker=a__ ) pipe.to(a__ ) pipe.set_progress_bar_config(disable=a__ ) pipe.enable_attention_slicing() __magic_name__ = '''Face of a yellow cat, high resolution, sitting on a park bench''' __magic_name__ = torch.manual_seed(0 ) __magic_name__ = pipe( prompt=a__ , image=a__ , mask_image=a__ , generator=a__ , output_type='''np''' , ) __magic_name__ = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 9E-3 def snake_case__ ( self : List[str] ): __magic_name__ = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''' ) __magic_name__ = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''' ) __magic_name__ = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint''' '''/yellow_cat_sitting_on_a_park_bench_fp16.npy''' ) __magic_name__ = '''stabilityai/stable-diffusion-2-inpainting''' __magic_name__ = StableDiffusionInpaintPipeline.from_pretrained( a__ , torch_dtype=torch.floataa , safety_checker=a__ , ) pipe.to(a__ ) pipe.set_progress_bar_config(disable=a__ ) pipe.enable_attention_slicing() __magic_name__ = '''Face of a yellow cat, high resolution, sitting on a park bench''' __magic_name__ = torch.manual_seed(0 ) __magic_name__ = pipe( prompt=a__ , image=a__ , mask_image=a__ , generator=a__ , output_type='''np''' , ) __magic_name__ = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 5E-1 def snake_case__ ( self : List[Any] ): torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __magic_name__ = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''' ) __magic_name__ = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''' ) __magic_name__ = '''stabilityai/stable-diffusion-2-inpainting''' __magic_name__ = PNDMScheduler.from_pretrained(a__ , subfolder='''scheduler''' ) __magic_name__ = StableDiffusionInpaintPipeline.from_pretrained( a__ , safety_checker=a__ , scheduler=a__ , torch_dtype=torch.floataa , ) pipe.to(a__ ) pipe.set_progress_bar_config(disable=a__ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() __magic_name__ = '''Face of a yellow cat, high resolution, sitting on a park bench''' __magic_name__ = torch.manual_seed(0 ) __magic_name__ = pipe( prompt=a__ , image=a__ , mask_image=a__ , generator=a__ , num_inference_steps=2 , output_type='''np''' , ) __magic_name__ = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 10**9
98
0
"""simple docstring""" import unittest from transformers import MPNetConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) class _UpperCAmelCase : def __init__( self :List[Any] , __UpperCamelCase :str , __UpperCamelCase :Tuple=13 , __UpperCamelCase :List[str]=7 , __UpperCamelCase :Optional[int]=True , __UpperCamelCase :List[str]=True , __UpperCamelCase :Union[str, Any]=False , __UpperCamelCase :Dict=True , __UpperCamelCase :str=99 , __UpperCamelCase :Any=64 , __UpperCamelCase :int=5 , __UpperCamelCase :Dict=4 , __UpperCamelCase :Tuple=64 , __UpperCamelCase :Union[str, Any]="gelu" , __UpperCamelCase :List[str]=0.1 , __UpperCamelCase :Optional[Any]=0.1 , __UpperCamelCase :List[Any]=5_12 , __UpperCamelCase :List[Any]=16 , __UpperCamelCase :str=2 , __UpperCamelCase :str=0.02 , __UpperCamelCase :Dict=3 , __UpperCamelCase :Optional[int]=4 , __UpperCamelCase :Dict=None , ): A = parent A = batch_size A = seq_length A = is_training A = use_input_mask A = use_token_type_ids A = use_labels A = vocab_size A = hidden_size A = num_hidden_layers A = num_attention_heads A = intermediate_size A = hidden_act A = hidden_dropout_prob A = attention_probs_dropout_prob A = max_position_embeddings A = type_vocab_size A = type_sequence_label_size A = initializer_range A = num_labels A = num_choices A = scope def lowerCamelCase ( self :str ): return MPNetConfig.from_pretrained("microsoft/mpnet-base" ) def lowerCamelCase ( self :Dict ): A = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A = None if self.use_input_mask: A = random_attention_mask([self.batch_size, self.seq_length] ) A = None A = None A = None if self.use_labels: A = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A = ids_tensor([self.batch_size] , self.num_choices ) A = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase ( self :Union[str, Any] ): return MPNetConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def lowerCamelCase ( self :str , __UpperCamelCase :Any , __UpperCamelCase :Dict , __UpperCamelCase :Union[str, Any] , __UpperCamelCase :Tuple , __UpperCamelCase :Union[str, Any] , __UpperCamelCase :List[str] ): A = MPNetModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() A = model(__UpperCamelCase , __UpperCamelCase ) A = model(__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def lowerCamelCase ( self :Optional[Any] , __UpperCamelCase :Any , __UpperCamelCase :Optional[Any] , __UpperCamelCase :Any , __UpperCamelCase :int , __UpperCamelCase :Dict , __UpperCamelCase :List[Any] ): A = MPNetForQuestionAnswering(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() A = model( __UpperCamelCase , attention_mask=__UpperCamelCase , start_positions=__UpperCamelCase , end_positions=__UpperCamelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase ( self :List[Any] , __UpperCamelCase :Dict , __UpperCamelCase :List[Any] , __UpperCamelCase :List[Any] , __UpperCamelCase :str , __UpperCamelCase :Tuple , __UpperCamelCase :List[Any] ): A = self.num_labels A = MPNetForSequenceClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() A = model(__UpperCamelCase , attention_mask=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase ( self :Dict , __UpperCamelCase :Optional[Any] , __UpperCamelCase :Any , __UpperCamelCase :List[Any] , __UpperCamelCase :Dict , __UpperCamelCase :Optional[Any] , __UpperCamelCase :int ): A = self.num_choices A = MPNetForMultipleChoice(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() A = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A = model( __UpperCamelCase , attention_mask=__UpperCamelCase , labels=__UpperCamelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCamelCase ( self :str , __UpperCamelCase :Optional[int] , __UpperCamelCase :int , __UpperCamelCase :Tuple , __UpperCamelCase :List[str] , __UpperCamelCase :Any , __UpperCamelCase :Optional[int] ): A = self.num_labels A = MPNetForTokenClassification(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() A = model(__UpperCamelCase , attention_mask=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCamelCase ( self :Optional[int] ): A = self.prepare_config_and_inputs() ((A), (A), (A), (A), (A), (A)) = config_and_inputs A = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class _UpperCAmelCase ( lowercase_ , lowercase_ , unittest.TestCase ): UpperCamelCase = ( ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) if is_torch_available() else () ) UpperCamelCase = ( { '''feature-extraction''': MPNetModel, '''fill-mask''': MPNetForMaskedLM, '''question-answering''': MPNetForQuestionAnswering, '''text-classification''': MPNetForSequenceClassification, '''token-classification''': MPNetForTokenClassification, '''zero-shot''': MPNetForSequenceClassification, } if is_torch_available() else {} ) UpperCamelCase = False UpperCamelCase = True def lowerCamelCase ( self :Tuple ): A = MPNetModelTester(self ) A = ConfigTester(self , config_class=__UpperCamelCase , hidden_size=37 ) def lowerCamelCase ( self :Optional[int] ): self.config_tester.run_common_tests() def lowerCamelCase ( self :Tuple ): A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_model(*__UpperCamelCase ) def lowerCamelCase ( self :Union[str, Any] ): A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_sequence_classification(*__UpperCamelCase ) def lowerCamelCase ( self :List[str] ): A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_multiple_choice(*__UpperCamelCase ) def lowerCamelCase ( self :int ): A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_token_classification(*__UpperCamelCase ) def lowerCamelCase ( self :str ): A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_question_answering(*__UpperCamelCase ) @require_torch class _UpperCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase ( self :List[Any] ): A = MPNetModel.from_pretrained("microsoft/mpnet-base" ) A = torch.tensor([[0, 3_45, 2_32, 3_28, 7_40, 1_40, 16_95, 69, 60_78, 15_88, 2]] ) A = model(__UpperCamelCase )[0] A = torch.Size((1, 11, 7_68) ) self.assertEqual(output.shape , __UpperCamelCase ) A = torch.tensor( [[[-0.0_550, 0.1_943, -0.0_740], [-0.0_562, 0.2_211, -0.0_579], [-0.0_437, 0.3_337, -0.0_641]]] ) # compare the actual values for a slice. self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCamelCase , atol=1e-4 ) )
292
"""simple docstring""" from typing import Callable, Dict, Optional, Tuple import torch from torch import nn from torch.distributions import ( AffineTransform, Distribution, Independent, NegativeBinomial, Normal, StudentT, TransformedDistribution, ) class _UpperCAmelCase ( lowercase_ ): def __init__( self :int , __UpperCamelCase :Distribution , __UpperCamelCase :Dict=None , __UpperCamelCase :Optional[int]=None , __UpperCamelCase :List[str]=0 ): A = 1.0 if scale is None else scale A = 0.0 if loc is None else loc super().__init__(__UpperCamelCase , [AffineTransform(loc=self.loc , scale=self.scale , event_dim=__UpperCamelCase )] ) @property def lowerCamelCase ( self :Any ): return self.base_dist.mean * self.scale + self.loc @property def lowerCamelCase ( self :Optional[int] ): return self.base_dist.variance * self.scale**2 @property def lowerCamelCase ( self :Dict ): return self.variance.sqrt() class _UpperCAmelCase ( nn.Module ): def __init__( self :Dict , __UpperCamelCase :int , __UpperCamelCase :Dict[str, int] , __UpperCamelCase :Callable[..., Tuple[torch.Tensor]] , **__UpperCamelCase :str ): super().__init__(**__UpperCamelCase ) A = args_dim A = nn.ModuleList([nn.Linear(__UpperCamelCase , __UpperCamelCase ) for dim in args_dim.values()] ) A = domain_map def lowerCamelCase ( self :int , __UpperCamelCase :torch.Tensor ): A = [proj(__UpperCamelCase ) for proj in self.proj] return self.domain_map(*__UpperCamelCase ) class _UpperCAmelCase ( nn.Module ): def __init__( self :Dict , __UpperCamelCase :int ): super().__init__() A = function def lowerCamelCase ( self :List[str] , __UpperCamelCase :Any , *__UpperCamelCase :Any ): return self.function(__UpperCamelCase , *__UpperCamelCase ) class _UpperCAmelCase : UpperCamelCase = 42 UpperCamelCase = 42 UpperCamelCase = 42 def __init__( self :Any , __UpperCamelCase :int = 1 ): A = dim A = {k: dim * self.args_dim[k] for k in self.args_dim} def lowerCamelCase ( self :List[Any] , __UpperCamelCase :Dict ): if self.dim == 1: return self.distribution_class(*__UpperCamelCase ) else: return Independent(self.distribution_class(*__UpperCamelCase ) , 1 ) def lowerCamelCase ( self :int , __UpperCamelCase :List[str] , __UpperCamelCase :Optional[torch.Tensor] = None , __UpperCamelCase :Optional[torch.Tensor] = None , ): A = self._base_distribution(__UpperCamelCase ) if loc is None and scale is None: return distr else: return AffineTransformed(__UpperCamelCase , loc=__UpperCamelCase , scale=__UpperCamelCase , event_dim=self.event_dim ) @property def lowerCamelCase ( self :List[Any] ): return () if self.dim == 1 else (self.dim,) @property def lowerCamelCase ( self :Tuple ): return len(self.event_shape ) @property def lowerCamelCase ( self :int ): return 0.0 def lowerCamelCase ( self :str , __UpperCamelCase :int ): return ParameterProjection( in_features=__UpperCamelCase , args_dim=self.args_dim , domain_map=LambdaLayer(self.domain_map ) , ) def lowerCamelCase ( self :List[Any] , *__UpperCamelCase :torch.Tensor ): raise NotImplementedError() @staticmethod def lowerCamelCase ( __UpperCamelCase :torch.Tensor ): return (x + torch.sqrt(torch.square(__UpperCamelCase ) + 4.0 )) / 2.0 class _UpperCAmelCase ( lowercase_ ): UpperCamelCase = {"df": 1, "loc": 1, "scale": 1} UpperCamelCase = StudentT @classmethod def lowerCamelCase ( cls :List[str] , __UpperCamelCase :torch.Tensor , __UpperCamelCase :torch.Tensor , __UpperCamelCase :torch.Tensor ): A = cls.squareplus(__UpperCamelCase ).clamp_min(torch.finfo(scale.dtype ).eps ) A = 2.0 + cls.squareplus(__UpperCamelCase ) return df.squeeze(-1 ), loc.squeeze(-1 ), scale.squeeze(-1 ) class _UpperCAmelCase ( lowercase_ ): UpperCamelCase = {"loc": 1, "scale": 1} UpperCamelCase = Normal @classmethod def lowerCamelCase ( cls :List[Any] , __UpperCamelCase :torch.Tensor , __UpperCamelCase :torch.Tensor ): A = cls.squareplus(__UpperCamelCase ).clamp_min(torch.finfo(scale.dtype ).eps ) return loc.squeeze(-1 ), scale.squeeze(-1 ) class _UpperCAmelCase ( lowercase_ ): UpperCamelCase = {"total_count": 1, "logits": 1} UpperCamelCase = NegativeBinomial @classmethod def lowerCamelCase ( cls :str , __UpperCamelCase :torch.Tensor , __UpperCamelCase :torch.Tensor ): A = cls.squareplus(__UpperCamelCase ) return total_count.squeeze(-1 ), logits.squeeze(-1 ) def lowerCamelCase ( self :Tuple , __UpperCamelCase :List[str] ): A, A = distr_args if self.dim == 1: return self.distribution_class(total_count=__UpperCamelCase , logits=__UpperCamelCase ) else: return Independent(self.distribution_class(total_count=__UpperCamelCase , logits=__UpperCamelCase ) , 1 ) def lowerCamelCase ( self :List[str] , __UpperCamelCase :str , __UpperCamelCase :Optional[torch.Tensor] = None , __UpperCamelCase :Optional[torch.Tensor] = None ): A, A = distr_args if scale is not None: # See scaling property of Gamma. logits += scale.log() return self._base_distribution((total_count, logits) )
292
1
from __future__ import annotations lowerCamelCase_ = list[list[int]] # assigning initial values to the grid lowerCamelCase_ = [ [3, 0, 6, 5, 0, 8, 4, 0, 0], [5, 2, 0, 0, 0, 0, 0, 0, 0], [0, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] # a grid with no solution lowerCamelCase_ = [ [5, 0, 6, 5, 0, 8, 4, 0, 3], [5, 2, 0, 0, 0, 0, 0, 0, 2], [1, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] def __magic_name__ ( __a : Matrix , __a : int , __a : int , __a : int ): '''simple docstring''' for i in range(9 ): if grid[row][i] == n or grid[i][column] == n: return False for i in range(3 ): for j in range(3 ): if grid[(row - row % 3) + i][(column - column % 3) + j] == n: return False return True def __magic_name__ ( __a : Matrix ): '''simple docstring''' for i in range(9 ): for j in range(9 ): if grid[i][j] == 0: return i, j return None def __magic_name__ ( __a : Matrix ): '''simple docstring''' if location := find_empty_location(__a ): UpperCamelCase__ , UpperCamelCase__ = location else: # If the location is ``None``, then the grid is solved. return grid for digit in range(1 , 10 ): if is_safe(__a , __a , __a , __a ): UpperCamelCase__ = digit if sudoku(__a ) is not None: return grid UpperCamelCase__ = 0 return None def __magic_name__ ( __a : Matrix ): '''simple docstring''' for row in grid: for cell in row: print(__a , end=""" """ ) print() if __name__ == "__main__": # make a copy of grid so that you can compare with the unmodified grid for example_grid in (initial_grid, no_solution): print('''\nExample grid:\n''' + '''=''' * 20) print_solution(example_grid) print('''\nExample grid solution:''') lowerCamelCase_ = sudoku(example_grid) if solution is not None: print_solution(solution) else: print('''Cannot find a solution.''')
352
import collections import os import re from pathlib import Path lowerCamelCase_ = '''src/transformers''' # Matches is_xxx_available() lowerCamelCase_ = re.compile(r'''is\_([a-z_]*)_available()''') # Catches a one-line _import_struct = {xxx} lowerCamelCase_ = re.compile(r'''^_import_structure\s+=\s+\{([^\}]+)\}''') # Catches a line with a key-values pattern: "bla": ["foo", "bar"] lowerCamelCase_ = re.compile(r'''\s+"\S*":\s+\[([^\]]*)\]''') # Catches a line if not is_foo_available lowerCamelCase_ = re.compile(r'''^\s*if\s+not\s+is\_[a-z_]*\_available\(\)''') # Catches a line _import_struct["bla"].append("foo") lowerCamelCase_ = re.compile(r'''^\s*_import_structure\["\S*"\]\.append\("(\S*)"\)''') # Catches a line _import_struct["bla"].extend(["foo", "bar"]) or _import_struct["bla"] = ["foo", "bar"] lowerCamelCase_ = re.compile(r'''^\s*_import_structure\[\S*\](?:\.extend\(|\s*=\s+)\[([^\]]*)\]''') # Catches a line with an object between quotes and a comma: "MyModel", lowerCamelCase_ = re.compile(r'''^\s+"([^"]+)",''') # Catches a line with objects between brackets only: ["foo", "bar"], lowerCamelCase_ = re.compile(r'''^\s+\[([^\]]+)\]''') # Catches a line with from foo import bar, bla, boo lowerCamelCase_ = re.compile(r'''\s+from\s+\S*\s+import\s+([^\(\s].*)\n''') # Catches a line with try: lowerCamelCase_ = re.compile(r'''^\s*try:''') # Catches a line with else: lowerCamelCase_ = re.compile(r'''^\s*else:''') def __magic_name__ ( __a : List[Any] ): '''simple docstring''' if _re_test_backend.search(__a ) is None: return None UpperCamelCase__ = [b[0] for b in _re_backend.findall(__a )] backends.sort() return "_and_".join(__a ) def __magic_name__ ( __a : Dict ): '''simple docstring''' with open(__a , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: UpperCamelCase__ = f.readlines() UpperCamelCase__ = 0 while line_index < len(__a ) and not lines[line_index].startswith("""_import_structure = {""" ): line_index += 1 # If this is a traditional init, just return. if line_index >= len(__a ): return None # First grab the objects without a specific backend in _import_structure UpperCamelCase__ = [] while not lines[line_index].startswith("""if TYPE_CHECKING""" ) and find_backend(lines[line_index] ) is None: UpperCamelCase__ = lines[line_index] # If we have everything on a single line, let's deal with it. if _re_one_line_import_struct.search(__a ): UpperCamelCase__ = _re_one_line_import_struct.search(__a ).groups()[0] UpperCamelCase__ = re.findall(R"""\[([^\]]+)\]""" , __a ) for imp in imports: objects.extend([obj[1:-1] for obj in imp.split(""", """ )] ) line_index += 1 continue UpperCamelCase__ = _re_import_struct_key_value.search(__a ) if single_line_import_search is not None: UpperCamelCase__ = [obj[1:-1] for obj in single_line_import_search.groups()[0].split(""", """ ) if len(__a ) > 0] objects.extend(__a ) elif line.startswith(""" """ * 8 + """\"""" ): objects.append(line[9:-3] ) line_index += 1 UpperCamelCase__ = {"""none""": objects} # Let's continue with backend-specific objects in _import_structure while not lines[line_index].startswith("""if TYPE_CHECKING""" ): # If the line is an if not is_backend_available, we grab all objects associated. UpperCamelCase__ = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: UpperCamelCase__ = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 UpperCamelCase__ = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(""" """ * 4 ): UpperCamelCase__ = lines[line_index] if _re_import_struct_add_one.search(__a ) is not None: objects.append(_re_import_struct_add_one.search(__a ).groups()[0] ) elif _re_import_struct_add_many.search(__a ) is not None: UpperCamelCase__ = _re_import_struct_add_many.search(__a ).groups()[0].split(""", """ ) UpperCamelCase__ = [obj[1:-1] for obj in imports if len(__a ) > 0] objects.extend(__a ) elif _re_between_brackets.search(__a ) is not None: UpperCamelCase__ = _re_between_brackets.search(__a ).groups()[0].split(""", """ ) UpperCamelCase__ = [obj[1:-1] for obj in imports if len(__a ) > 0] objects.extend(__a ) elif _re_quote_object.search(__a ) is not None: objects.append(_re_quote_object.search(__a ).groups()[0] ) elif line.startswith(""" """ * 8 + """\"""" ): objects.append(line[9:-3] ) elif line.startswith(""" """ * 12 + """\"""" ): objects.append(line[13:-3] ) line_index += 1 UpperCamelCase__ = objects else: line_index += 1 # At this stage we are in the TYPE_CHECKING part, first grab the objects without a specific backend UpperCamelCase__ = [] while ( line_index < len(__a ) and find_backend(lines[line_index] ) is None and not lines[line_index].startswith("""else""" ) ): UpperCamelCase__ = lines[line_index] UpperCamelCase__ = _re_import.search(__a ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(""", """ ) ) elif line.startswith(""" """ * 8 ): objects.append(line[8:-2] ) line_index += 1 UpperCamelCase__ = {"""none""": objects} # Let's continue with backend-specific objects while line_index < len(__a ): # If the line is an if is_backend_available, we grab all objects associated. UpperCamelCase__ = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: UpperCamelCase__ = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 UpperCamelCase__ = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(""" """ * 8 ): UpperCamelCase__ = lines[line_index] UpperCamelCase__ = _re_import.search(__a ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(""", """ ) ) elif line.startswith(""" """ * 12 ): objects.append(line[12:-2] ) line_index += 1 UpperCamelCase__ = objects else: line_index += 1 return import_dict_objects, type_hint_objects def __magic_name__ ( __a : Dict , __a : Dict ): '''simple docstring''' def find_duplicates(__a : Optional[Any] ): return [k for k, v in collections.Counter(__a ).items() if v > 1] if list(import_dict_objects.keys() ) != list(type_hint_objects.keys() ): return ["Both sides of the init do not have the same backends!"] UpperCamelCase__ = [] for key in import_dict_objects.keys(): UpperCamelCase__ = find_duplicates(import_dict_objects[key] ) if duplicate_imports: errors.append(f"Duplicate _import_structure definitions for: {duplicate_imports}" ) UpperCamelCase__ = find_duplicates(type_hint_objects[key] ) if duplicate_type_hints: errors.append(f"Duplicate TYPE_CHECKING objects for: {duplicate_type_hints}" ) if sorted(set(import_dict_objects[key] ) ) != sorted(set(type_hint_objects[key] ) ): UpperCamelCase__ = """base imports""" if key == """none""" else f"{key} backend" errors.append(f"Differences for {name}:" ) for a in type_hint_objects[key]: if a not in import_dict_objects[key]: errors.append(f" {a} in TYPE_HINT but not in _import_structure." ) for a in import_dict_objects[key]: if a not in type_hint_objects[key]: errors.append(f" {a} in _import_structure but not in TYPE_HINT." ) return errors def __magic_name__ ( ): '''simple docstring''' UpperCamelCase__ = [] for root, _, files in os.walk(__a ): if "__init__.py" in files: UpperCamelCase__ = os.path.join(__a , """__init__.py""" ) UpperCamelCase__ = parse_init(__a ) if objects is not None: UpperCamelCase__ = analyze_results(*__a ) if len(__a ) > 0: UpperCamelCase__ = f"Problem in {fname}, both halves do not define the same objects.\n{errors[0]}" failures.append("""\n""".join(__a ) ) if len(__a ) > 0: raise ValueError("""\n\n""".join(__a ) ) def __magic_name__ ( ): '''simple docstring''' UpperCamelCase__ = [] for path, directories, files in os.walk(__a ): for folder in directories: # Ignore private modules if folder.startswith("""_""" ): directories.remove(__a ) continue # Ignore leftovers from branches (empty folders apart from pycache) if len(list((Path(__a ) / folder).glob("""*.py""" ) ) ) == 0: continue UpperCamelCase__ = str((Path(__a ) / folder).relative_to(__a ) ) UpperCamelCase__ = short_path.replace(os.path.sep , """.""" ) submodules.append(__a ) for fname in files: if fname == "__init__.py": continue UpperCamelCase__ = str((Path(__a ) / fname).relative_to(__a ) ) UpperCamelCase__ = short_path.replace(""".py""" , """""" ).replace(os.path.sep , """.""" ) if len(submodule.split(""".""" ) ) == 1: submodules.append(__a ) return submodules lowerCamelCase_ = [ '''convert_pytorch_checkpoint_to_tf2''', '''modeling_flax_pytorch_utils''', '''models.esm.openfold_utils''', ] def __magic_name__ ( ): '''simple docstring''' from transformers.utils import direct_transformers_import UpperCamelCase__ = direct_transformers_import(__a ) UpperCamelCase__ = set(transformers._import_structure.keys() ) # This contains all the base keys of the _import_structure object defined in the init, but if the user is missing # some optional dependencies, they may not have all of them. Thus we read the init to read all additions and # (potentiall re-) add them. with open(os.path.join(__a , """__init__.py""" ) , """r""" ) as f: UpperCamelCase__ = f.read() import_structure_keys.update(set(re.findall(R"""import_structure\[\"([^\"]*)\"\]""" , __a ) ) ) UpperCamelCase__ = [ module for module in get_transformers_submodules() if module not in IGNORE_SUBMODULES and module not in import_structure_keys ] if len(__a ) > 0: UpperCamelCase__ = """\n""".join(f"- {module}" for module in module_not_registered ) raise ValueError( """The following submodules are not properly registed in the main init of Transformers:\n""" f"{list_of_modules}\n" """Make sure they appear somewhere in the keys of `_import_structure` with an empty list as value.""" ) if __name__ == "__main__": check_all_inits() check_submodules()
178
0
from itertools import zip_longest import requests from bsa import BeautifulSoup from pandas import DataFrame def lowercase_ ( _lowerCamelCase : str = "laptop"): lowercase__ : Optional[Any] = f'''https://www.amazon.in/laptop/s?k={product}''' lowercase__ : Dict = { "User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36\n (KHTML, like Gecko)Chrome/44.0.2403.157 Safari/537.36", "Accept-Language": "en-US, en;q=0.5", } lowercase__ : List[str] = BeautifulSoup(requests.get(_lowerCamelCase , headers=_lowerCamelCase).text) # Initialize a Pandas dataframe with the column titles lowercase__ : Dict = DataFrame( columns=[ "Product Title", "Product Link", "Current Price of the product", "Product Rating", "MRP of the product", "Discount", ]) # Loop through each entry and store them in the dataframe for item, _ in zip_longest( soup.find_all( "div" , attrs={"class": "s-result-item", "data-component-type": "s-search-result"} , ) , soup.find_all("div" , attrs={"class": "a-row a-size-base a-color-base"}) , ): try: lowercase__ : List[str] = item.ha.text lowercase__ : List[str] = "https://www.amazon.in/" + item.ha.a["href"] lowercase__ : Optional[Any] = item.find("span" , attrs={"class": "a-offscreen"}).text try: lowercase__ : Tuple = item.find("span" , attrs={"class": "a-icon-alt"}).text except AttributeError: lowercase__ : List[Any] = "Not available" try: lowercase__ : int = ( "₹" + item.find( "span" , attrs={"class": "a-price a-text-price"}).text.split("₹")[1] ) except AttributeError: lowercase__ : Any = "" try: lowercase__ : Optional[Any] = float( ( ( float(product_mrp.strip("₹").replace("," , "")) - float(product_price.strip("₹").replace("," , "")) ) / float(product_mrp.strip("₹").replace("," , "")) ) * 100) except ValueError: lowercase__ : Dict = float("nan") except AttributeError: pass lowercase__ : Tuple = [ product_title, product_link, product_price, product_rating, product_mrp, discount, ] lowercase__ : str = " " lowercase__ : Dict = " " data_frame.index += 1 return data_frame if __name__ == "__main__": UpperCamelCase = '''headphones''' get_amazon_product_data(product).to_csv(f"Amazon Product Data for {product}.csv")
87
def A__ ( __lowerCamelCase ): SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = { '''^''': 3, '''*''': 2, '''/''': 2, '''%''': 2, '''+''': 1, '''-''': 1, } # Priority of each operator SCREAMING_SNAKE_CASE_ = len(__lowerCamelCase ) if (len(__lowerCamelCase ) > 7) else 7 # Print table header for output print( '''Symbol'''.center(8 ), '''Stack'''.center(__lowerCamelCase ), '''Postfix'''.center(__lowerCamelCase ), sep=''' | ''', ) print('''-''' * (print_width * 3 + 7) ) for x in infix: if x.isalpha() or x.isdigit(): post_fix.append(__lowerCamelCase ) # if x is Alphabet / Digit, add it to Postfix elif x == "(": stack.append(__lowerCamelCase ) # if x is "(" push to Stack elif x == ")": # if x is ")" pop stack until "(" is encountered while stack[-1] != "(": post_fix.append(stack.pop() ) # Pop stack & add the content to Postfix stack.pop() else: if len(__lowerCamelCase ) == 0: stack.append(__lowerCamelCase ) # If stack is empty, push x to stack else: # while priority of x is not > priority of element in the stack while len(__lowerCamelCase ) > 0 and priority[x] <= priority[stack[-1]]: post_fix.append(stack.pop() ) # pop stack & add to Postfix stack.append(__lowerCamelCase ) # push x to stack print( x.center(8 ), (''''''.join(__lowerCamelCase )).ljust(__lowerCamelCase ), (''''''.join(__lowerCamelCase )).ljust(__lowerCamelCase ), sep=''' | ''', ) # Output in tabular format while len(__lowerCamelCase ) > 0: # while stack is not empty post_fix.append(stack.pop() ) # pop stack & add to Postfix print( ''' '''.center(8 ), (''''''.join(__lowerCamelCase )).ljust(__lowerCamelCase ), (''''''.join(__lowerCamelCase )).ljust(__lowerCamelCase ), sep=''' | ''', ) # Output in tabular format return "".join(__lowerCamelCase ) # return Postfix as str def A__ ( __lowerCamelCase ): SCREAMING_SNAKE_CASE_ = list(infix[::-1] ) # reverse the infix equation for i in range(len(__lowerCamelCase ) ): if infix[i] == "(": SCREAMING_SNAKE_CASE_ = ''')''' # change "(" to ")" elif infix[i] == ")": SCREAMING_SNAKE_CASE_ = '''(''' # change ")" to "(" return (infix_2_postfix(''''''.join(__lowerCamelCase ) ))[ ::-1 ] # call infix_2_postfix on Infix, return reverse of Postfix if __name__ == "__main__": __UpperCAmelCase = input("\nEnter an Infix Equation = ") # Input an Infix equation __UpperCAmelCase = "".join(Infix.split()) # Remove spaces from the input print("\n\t", Infix, "(Infix) -> ", infix_2_prefix(Infix), "(Prefix)")
299
0
'''simple docstring''' import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class lowerCAmelCase__ ( lowerCamelCase_ ): lowerCAmelCase_ = ['''image_processor''', '''tokenizer'''] lowerCAmelCase_ = '''LayoutLMv2ImageProcessor''' lowerCAmelCase_ = ('''LayoutXLMTokenizer''', '''LayoutXLMTokenizerFast''') def __init__( self , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None , **__SCREAMING_SNAKE_CASE ): """simple docstring""" if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , __SCREAMING_SNAKE_CASE , ) lowercase_ : Optional[int] = kwargs.pop('''feature_extractor''' ) lowercase_ : Dict = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) def __call__( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = True , __SCREAMING_SNAKE_CASE = False , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = 0 , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = False , __SCREAMING_SNAKE_CASE = False , __SCREAMING_SNAKE_CASE = False , __SCREAMING_SNAKE_CASE = False , __SCREAMING_SNAKE_CASE = True , __SCREAMING_SNAKE_CASE = None , **__SCREAMING_SNAKE_CASE , ): """simple docstring""" if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( '''You cannot provide bounding boxes ''' '''if you initialized the image processor with apply_ocr set to True.''' ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( '''You cannot provide word labels if you initialized the image processor with apply_ocr set to True.''' ) if return_overflowing_tokens is True and return_offsets_mapping is False: raise ValueError('''You cannot return overflowing tokens without returning the offsets mapping.''' ) # first, apply the image processor lowercase_ : Union[str, Any] = self.image_processor(images=__SCREAMING_SNAKE_CASE , return_tensors=__SCREAMING_SNAKE_CASE ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowercase_ : Dict = [text] # add batch dimension (as the image processor always adds a batch dimension) lowercase_ : Optional[int] = features['''words'''] lowercase_ : Dict = self.tokenizer( text=text if text is not None else features['''words'''] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features['''boxes'''] , word_labels=__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE , padding=__SCREAMING_SNAKE_CASE , truncation=__SCREAMING_SNAKE_CASE , max_length=__SCREAMING_SNAKE_CASE , stride=__SCREAMING_SNAKE_CASE , pad_to_multiple_of=__SCREAMING_SNAKE_CASE , return_token_type_ids=__SCREAMING_SNAKE_CASE , return_attention_mask=__SCREAMING_SNAKE_CASE , return_overflowing_tokens=__SCREAMING_SNAKE_CASE , return_special_tokens_mask=__SCREAMING_SNAKE_CASE , return_offsets_mapping=__SCREAMING_SNAKE_CASE , return_length=__SCREAMING_SNAKE_CASE , verbose=__SCREAMING_SNAKE_CASE , return_tensors=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE , ) # add pixel values lowercase_ : List[Any] = features.pop('''pixel_values''' ) if return_overflowing_tokens is True: lowercase_ : Dict = self.get_overflowing_images(__SCREAMING_SNAKE_CASE , encoded_inputs['''overflow_to_sample_mapping'''] ) lowercase_ : int = images return encoded_inputs def _snake_case ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase_ : Optional[int] = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(__SCREAMING_SNAKE_CASE ) != len(__SCREAMING_SNAKE_CASE ): raise ValueError( '''Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got''' F''' {len(__SCREAMING_SNAKE_CASE )} and {len(__SCREAMING_SNAKE_CASE )}''' ) return images_with_overflow def _snake_case ( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" return self.tokenizer.batch_decode(*__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) def _snake_case ( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" return self.tokenizer.decode(*__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) @property def _snake_case ( self ): """simple docstring""" return ["input_ids", "bbox", "attention_mask", "image"] @property def _snake_case ( self ): """simple docstring""" warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , __SCREAMING_SNAKE_CASE , ) return self.image_processor_class @property def _snake_case ( self ): """simple docstring""" warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , __SCREAMING_SNAKE_CASE , ) return self.image_processor
264
'''simple docstring''' from maths.is_square_free import is_square_free from maths.prime_factors import prime_factors def snake_case_ ( __SCREAMING_SNAKE_CASE : int ): """simple docstring""" lowercase_ : Tuple = prime_factors(__SCREAMING_SNAKE_CASE ) if is_square_free(__SCREAMING_SNAKE_CASE ): return -1 if len(__SCREAMING_SNAKE_CASE ) % 2 else 1 return 0 if __name__ == "__main__": import doctest doctest.testmod()
264
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _a : str= logging.get_logger(__name__) _a : Optional[int]= { "microsoft/focalnet-tiny": "https://huggingface.co/microsoft/focalnet-tiny/resolve/main/config.json", } class UpperCamelCase ( lowercase , lowercase ): UpperCAmelCase : Union[str, Any] = """focalnet""" def __init__(self : Union[str, Any] , _A : Any=2_24 , _A : Tuple=4 , _A : int=3 , _A : str=96 , _A : Tuple=False , _A : Any=[1_92, 3_84, 7_68, 7_68] , _A : Any=[2, 2, 6, 2] , _A : Optional[Any]=[2, 2, 2, 2] , _A : List[Any]=[3, 3, 3, 3] , _A : Union[str, Any]="gelu" , _A : List[str]=4.0 , _A : List[Any]=0.0 , _A : Optional[int]=0.1 , _A : Optional[Any]=False , _A : Tuple=1E-4 , _A : Union[str, Any]=False , _A : List[str]=False , _A : Optional[Any]=False , _A : str=0.02 , _A : Union[str, Any]=1E-5 , _A : str=32 , _A : str=None , _A : int=None , **_A : Optional[int] , ) -> Optional[int]: super().__init__(**_A) __snake_case : Union[str, Any] = image_size __snake_case : str = patch_size __snake_case : Union[str, Any] = num_channels __snake_case : Tuple = embed_dim __snake_case : Tuple = use_conv_embed __snake_case : List[str] = hidden_sizes __snake_case : List[str] = depths __snake_case : Tuple = focal_levels __snake_case : str = focal_windows __snake_case : int = hidden_act __snake_case : Union[str, Any] = mlp_ratio __snake_case : Union[str, Any] = hidden_dropout_prob __snake_case : Dict = drop_path_rate __snake_case : List[str] = use_layerscale __snake_case : Any = layerscale_value __snake_case : List[str] = use_post_layernorm __snake_case : Tuple = use_post_layernorm_in_modulation __snake_case : Optional[Any] = normalize_modulator __snake_case : Optional[Any] = initializer_range __snake_case : int = layer_norm_eps __snake_case : Tuple = encoder_stride __snake_case : int = ['stem'] + [f"stage{idx}" for idx in range(1 , len(self.depths) + 1)] __snake_case , __snake_case : Union[str, Any] = get_aligned_output_features_output_indices( out_features=_A , out_indices=_A , stage_names=self.stage_names)
172
"""simple docstring""" import os from argparse import ArgumentParser, Namespace from ..data import SingleSentenceClassificationProcessor as Processor from ..pipelines import TextClassificationPipeline from ..utils import is_tf_available, is_torch_available, logging from . import BaseTransformersCLICommand if not is_tf_available() and not is_torch_available(): raise RuntimeError("At least one of PyTorch or TensorFlow 2.0+ should be installed to use CLI training") # TF training parameters _a : Optional[int]= False _a : int= False def __UpperCAmelCase ( UpperCAmelCase_ : Namespace ) -> Optional[Any]: '''simple docstring''' return TrainCommand(UpperCAmelCase_ ) class UpperCamelCase ( lowercase ): @staticmethod def _lowercase (_A : ArgumentParser) -> Any: __snake_case : Any = parser.add_parser('train' , help='CLI tool to train a model on a task.') train_parser.add_argument( '--train_data' , type=_A , required=_A , help='path to train (and optionally evaluation) dataset as a csv with tab separated labels and sentences.' , ) train_parser.add_argument( '--column_label' , type=_A , default=0 , help='Column of the dataset csv file with example labels.') train_parser.add_argument( '--column_text' , type=_A , default=1 , help='Column of the dataset csv file with example texts.') train_parser.add_argument( '--column_id' , type=_A , default=2 , help='Column of the dataset csv file with example ids.') train_parser.add_argument( '--skip_first_row' , action='store_true' , help='Skip the first row of the csv file (headers).') train_parser.add_argument('--validation_data' , type=_A , default='' , help='path to validation dataset.') train_parser.add_argument( '--validation_split' , type=_A , default=0.1 , help='if validation dataset is not provided, fraction of train dataset to use as validation dataset.' , ) train_parser.add_argument('--output' , type=_A , default='./' , help='path to saved the trained model.') train_parser.add_argument( '--task' , type=_A , default='text_classification' , help='Task to train the model on.') train_parser.add_argument( '--model' , type=_A , default='bert-base-uncased' , help='Model\'s name or path to stored model.') train_parser.add_argument('--train_batch_size' , type=_A , default=32 , help='Batch size for training.') train_parser.add_argument('--valid_batch_size' , type=_A , default=64 , help='Batch size for validation.') train_parser.add_argument('--learning_rate' , type=_A , default=3E-5 , help='Learning rate.') train_parser.add_argument('--adam_epsilon' , type=_A , default=1E-08 , help='Epsilon for Adam optimizer.') train_parser.set_defaults(func=_A) def __init__(self : int , _A : Namespace) -> Tuple: __snake_case : Optional[int] = logging.get_logger('transformers-cli/training') __snake_case : Optional[int] = 'tf' if is_tf_available() else 'torch' os.makedirs(args.output , exist_ok=_A) __snake_case : List[Any] = args.output __snake_case : Any = args.column_label __snake_case : str = args.column_text __snake_case : Any = args.column_id self.logger.info(f"Loading {args.task} pipeline for {args.model}") if args.task == "text_classification": __snake_case : List[str] = TextClassificationPipeline.from_pretrained(args.model) elif args.task == "token_classification": raise NotImplementedError elif args.task == "question_answering": raise NotImplementedError self.logger.info(f"Loading dataset from {args.train_data}") __snake_case : List[Any] = Processor.create_from_csv( args.train_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) __snake_case : List[str] = None if args.validation_data: self.logger.info(f"Loading validation dataset from {args.validation_data}") __snake_case : Dict = Processor.create_from_csv( args.validation_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) __snake_case : List[str] = args.validation_split __snake_case : str = args.train_batch_size __snake_case : Any = args.valid_batch_size __snake_case : Union[str, Any] = args.learning_rate __snake_case : str = args.adam_epsilon def _lowercase (self : List[str]) -> str: if self.framework == "tf": return self.run_tf() return self.run_torch() def _lowercase (self : str) -> int: raise NotImplementedError def _lowercase (self : Union[str, Any]) -> Optional[Any]: self.pipeline.fit( self.train_dataset , validation_data=self.valid_dataset , validation_split=self.validation_split , learning_rate=self.learning_rate , adam_epsilon=self.adam_epsilon , train_batch_size=self.train_batch_size , valid_batch_size=self.valid_batch_size , ) # Save trained pipeline self.pipeline.save_pretrained(self.output)
172
1
'''simple docstring''' import warnings from ...utils import logging from .image_processing_segformer import SegformerImageProcessor UpperCamelCase = logging.get_logger(__name__) class lowerCAmelCase_ ( UpperCAmelCase_ ): '''simple docstring''' def __init__( self : int , *SCREAMING_SNAKE_CASE_ : str , **SCREAMING_SNAKE_CASE_ : Dict ) -> None: '''simple docstring''' warnings.warn( '''The class SegformerFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use SegformerImageProcessor instead.''' , SCREAMING_SNAKE_CASE_ , ) super().__init__(*SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ )
362
'''simple docstring''' from ..utils import is_flax_available, is_torch_available if is_torch_available(): from .autoencoder_kl import AutoencoderKL from .controlnet import ControlNetModel from .dual_transformer_ad import DualTransformeraDModel from .modeling_utils import ModelMixin from .prior_transformer import PriorTransformer from .ta_film_transformer import TaFilmDecoder from .transformer_ad import TransformeraDModel from .unet_ad import UNetaDModel from .unet_ad import UNetaDModel from .unet_ad_condition import UNetaDConditionModel from .unet_ad_condition import UNetaDConditionModel from .vq_model import VQModel if is_flax_available(): from .controlnet_flax import FlaxControlNetModel from .unet_ad_condition_flax import FlaxUNetaDConditionModel from .vae_flax import FlaxAutoencoderKL
334
0
'''simple docstring''' import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class __A ( unittest.TestCase ): def _lowercase (self : Any ): UpperCAmelCase_ = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(__a ) ) def _lowercase (self : int ): UpperCAmelCase_ = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertTrue(is_safetensors_compatible(__a ) ) def _lowercase (self : Optional[Any] ): UpperCAmelCase_ = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", "unet/diffusion_pytorch_model.bin", # Removed: 'unet/diffusion_pytorch_model.safetensors', ] self.assertFalse(is_safetensors_compatible(__a ) ) def _lowercase (self : Union[str, Any] ): UpperCAmelCase_ = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] self.assertTrue(is_safetensors_compatible(__a ) ) def _lowercase (self : str ): UpperCAmelCase_ = [ "safety_checker/pytorch_model.bin", "safety_checker/model.safetensors", "vae/diffusion_pytorch_model.bin", "vae/diffusion_pytorch_model.safetensors", "text_encoder/pytorch_model.bin", # Removed: 'text_encoder/model.safetensors', "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] self.assertFalse(is_safetensors_compatible(__a ) ) def _lowercase (self : Dict ): UpperCAmelCase_ = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] UpperCAmelCase_ = "fp16" self.assertTrue(is_safetensors_compatible(__a , variant=__a ) ) def _lowercase (self : Optional[Any] ): UpperCAmelCase_ = [ "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] UpperCAmelCase_ = "fp16" self.assertTrue(is_safetensors_compatible(__a , variant=__a ) ) def _lowercase (self : Optional[int] ): # pass variant but use the non-variant filenames UpperCAmelCase_ = [ "unet/diffusion_pytorch_model.bin", "unet/diffusion_pytorch_model.safetensors", ] UpperCAmelCase_ = "fp16" self.assertTrue(is_safetensors_compatible(__a , variant=__a ) ) def _lowercase (self : Union[str, Any] ): UpperCAmelCase_ = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", "unet/diffusion_pytorch_model.fp16.bin", # Removed: 'unet/diffusion_pytorch_model.fp16.safetensors', ] UpperCAmelCase_ = "fp16" self.assertFalse(is_safetensors_compatible(__a , variant=__a ) ) def _lowercase (self : str ): UpperCAmelCase_ = [ "text_encoder/pytorch_model.fp16.bin", "text_encoder/model.fp16.safetensors", ] UpperCAmelCase_ = "fp16" self.assertTrue(is_safetensors_compatible(__a , variant=__a ) ) def _lowercase (self : Union[str, Any] ): # pass variant but use the non-variant filenames UpperCAmelCase_ = [ "text_encoder/pytorch_model.bin", "text_encoder/model.safetensors", ] UpperCAmelCase_ = "fp16" self.assertTrue(is_safetensors_compatible(__a , variant=__a ) ) def _lowercase (self : int ): UpperCAmelCase_ = [ "safety_checker/pytorch_model.fp16.bin", "safety_checker/model.fp16.safetensors", "vae/diffusion_pytorch_model.fp16.bin", "vae/diffusion_pytorch_model.fp16.safetensors", "text_encoder/pytorch_model.fp16.bin", # 'text_encoder/model.fp16.safetensors', "unet/diffusion_pytorch_model.fp16.bin", "unet/diffusion_pytorch_model.fp16.safetensors", ] UpperCAmelCase_ = "fp16" self.assertFalse(is_safetensors_compatible(__a , variant=__a ) )
1
import argparse import re from flax.traverse_util import flatten_dict, unflatten_dict from tax import checkpoints from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model from transformers.utils import logging logging.set_verbosity_info() # should not include what is already done by the `from_pt` argument __lowerCamelCase : Any = { '''/attention/''': '''/0/SelfAttention/''', '''/self_attention/''': '''/0/SelfAttention/''', '''/encoder_decoder_attention/''': '''/1/EncDecAttention/''', '''value''': '''v''', '''query''': '''q''', '''key''': '''k''', '''out''': '''o''', '''pre_self_attention_layer_norm''': '''0/layer_norm''', '''pre_cross_attention_layer_norm''': '''1/layer_norm''', '''pre_attention_layer_norm''': '''0/layer_norm''', # previously 1, but seems wrong '''token_embedder''': '''shared''', '''encoder_norm''': '''final_layer_norm''', '''decoder_norm''': '''final_layer_norm''', '''relpos_bias/rel_embedding''': '''block/0/layer/0/SelfAttention/relative_attention_bias/weight''', '''router/router_weights/w/''': '''router/classifier/''', '''roer/roer_weights/w/''': '''router/classifier/''', '''logits_dense''': '''lm_head''', } def _snake_case ( lowerCAmelCase : List[str] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = list(s_dict.keys() ) for key in keys: SCREAMING_SNAKE_CASE_ : int = R".*/layers_(\d+)" SCREAMING_SNAKE_CASE_ : List[Any] = key if re.match(lowerCAmelCase , lowerCAmelCase ): SCREAMING_SNAKE_CASE_ : Any = re.sub(R"layers_(\d+)" , R"block/\1/layer" , lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : List[Any] = R"(encoder|decoder)\/" if re.match(lowerCAmelCase , lowerCAmelCase ): SCREAMING_SNAKE_CASE_ : str = re.match(lowerCAmelCase , lowerCAmelCase ).groups() if groups[0] == "encoder": SCREAMING_SNAKE_CASE_ : Any = re.sub(R"/mlp/" , R"/1/mlp/" , lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : List[Any] = re.sub(R"/pre_mlp_layer_norm/" , R"/1/layer_norm/" , lowerCAmelCase ) elif groups[0] == "decoder": SCREAMING_SNAKE_CASE_ : List[str] = re.sub(R"/mlp/" , R"/2/mlp/" , lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Dict = re.sub(R"/pre_mlp_layer_norm/" , R"/2/layer_norm/" , lowerCAmelCase ) # 2. Convert other classic mappings for old_key, temp_key in MOE_LAYER_NAME_MAPPING.items(): if old_key in new_key: SCREAMING_SNAKE_CASE_ : List[Any] = new_key.replace(lowerCAmelCase , lowerCAmelCase ) print(f'{key} -> {new_key}' ) SCREAMING_SNAKE_CASE_ : List[Any] = s_dict.pop(lowerCAmelCase ) if "encoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" in s_dict: SCREAMING_SNAKE_CASE_ : str = s_dict[ "encoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" ].T if "decoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" in s_dict: SCREAMING_SNAKE_CASE_ : Optional[int] = s_dict[ "decoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" ].T # 3. Take extra care of the EXPERTS layer for key in list(s_dict.keys() ): if "expert" in key: SCREAMING_SNAKE_CASE_ : Union[str, Any] = s_dict[key].shape[0] SCREAMING_SNAKE_CASE_ : List[Any] = s_dict[key] for idx in range(lowerCAmelCase ): SCREAMING_SNAKE_CASE_ : Tuple = expert_weihts[idx] print(f'{key} -> {key.replace("expert/" , "nested fstring" )}' ) s_dict.pop(lowerCAmelCase ) return s_dict __lowerCamelCase : List[Any] = { '''NUM_ENCODER_LAYERS''': '''num_layers''', '''NUM_DECODER_LAYERS''': '''num_decoder_layers''', '''NUM_HEADS''': '''num_heads''', '''HEAD_DIM''': '''d_kv''', '''EMBED_DIM''': '''d_model''', '''MLP_DIM''': '''d_ff''', '''NUM_SELECTED_EXPERTS''': '''num_selected_experts''', '''NUM_ENCODER_SPARSE_LAYERS''': '''num_sparse_encoder_layers''', '''NUM_DECODER_SPARSE_LAYERS''': '''num_sparse_decoder_layers''', '''dense.MlpBlock.activations''': '''feed_forward_proj''', } def _snake_case ( lowerCAmelCase : Dict , lowerCAmelCase : Optional[Any] ): """simple docstring""" import regex as re with open(lowerCAmelCase , "r" ) as f: SCREAMING_SNAKE_CASE_ : Optional[Any] = f.read() SCREAMING_SNAKE_CASE_ : List[str] = re.findall(R"(.*) = ([0-9.]*)" , lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Any = {} for param, value in regex_match: if param in GIN_TO_CONFIG_MAPPING and value != "": SCREAMING_SNAKE_CASE_ : int = float(lowerCAmelCase ) if "." in value else int(lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = re.findall(R"(.*activations) = \(\'(.*)\',\)" , lowerCAmelCase )[0] SCREAMING_SNAKE_CASE_ : List[str] = str(activation[1] ) SCREAMING_SNAKE_CASE_ : str = num_experts SCREAMING_SNAKE_CASE_ : Tuple = SwitchTransformersConfig(**lowerCAmelCase ) return config def _snake_case ( lowerCAmelCase : str , lowerCAmelCase : str , lowerCAmelCase : str=None , lowerCAmelCase : Optional[Any]="./" , lowerCAmelCase : Dict=8 ): """simple docstring""" print(f'Loading flax weights from : {flax_checkpoint_path}' ) SCREAMING_SNAKE_CASE_ : int = checkpoints.load_tax_checkpoint(lowerCAmelCase ) if gin_file is not None: SCREAMING_SNAKE_CASE_ : int = convert_gin_to_config(lowerCAmelCase , lowerCAmelCase ) else: SCREAMING_SNAKE_CASE_ : Dict = SwitchTransformersConfig.from_pretrained(lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : str = SwitchTransformersForConditionalGeneration(lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Optional[int] = flax_params["target"] SCREAMING_SNAKE_CASE_ : List[str] = flatten_dict(lowerCAmelCase , sep="/" ) SCREAMING_SNAKE_CASE_ : List[str] = rename_keys(lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Optional[int] = unflatten_dict(lowerCAmelCase , sep="/" ) # Load the flax params in the PT model load_flax_weights_in_pytorch_model(lowerCAmelCase , lowerCAmelCase ) print(f'Save PyTorch model to {pytorch_dump_path}' ) pt_model.save_pretrained(lowerCAmelCase ) if __name__ == "__main__": __lowerCamelCase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--switch_t5x_checkpoint_path''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained SwitchTransformers model. \nThis specifies the''' ''' model architecture. If not provided, a `gin_file` has to be provided.''' ), ) parser.add_argument( '''--gin_file''', default=None, type=str, required=False, help='''Path to the gin config file. If not provided, a `config_file` has to be passed ''', ) parser.add_argument( '''--config_name''', default=None, type=str, required=False, help='''Config name of SwitchTransformers model.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output pytorch model.''' ) parser.add_argument('''--num_experts''', default=8, type=int, required=False, help='''Number of experts''') __lowerCamelCase : Any = parser.parse_args() convert_flax_checkpoint_to_pytorch( args.switch_tax_checkpoint_path, args.config_name, args.gin_file, args.pytorch_dump_folder_path, args.num_experts, )
18
0
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): lowercase__ = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): lowercase__ = n - k # Calculate C(n,k) for i in range(SCREAMING_SNAKE_CASE_ ): result *= n - i result //= i + 1 return result def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): return binomial_coefficient(2 * node_count , SCREAMING_SNAKE_CASE_ ) // (node_count + 1) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): if n < 0: raise ValueError("factorial() not defined for negative values" ) lowercase__ = 1 for i in range(1 , n + 1 ): result *= i return result def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): return catalan_number(SCREAMING_SNAKE_CASE_ ) * factorial(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": lowercase_ = int(input("""Enter the number of nodes: """).strip() or 0) if node_count <= 0: raise ValueError("""We need some nodes to work with.""") print( F'Given {node_count} nodes, there are {binary_tree_count(node_count)} ' F'binary trees and {catalan_number(node_count)} binary search trees.' )
224
def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = int(SCREAMING_SNAKE_CASE_ ) if decimal in (0, 1): # Exit cases for the recursion return str(SCREAMING_SNAKE_CASE_ ) lowercase__ , lowercase__ = divmod(SCREAMING_SNAKE_CASE_ , 2 ) return binary_recursive(SCREAMING_SNAKE_CASE_ ) + str(SCREAMING_SNAKE_CASE_ ) def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ ): lowercase__ = str(SCREAMING_SNAKE_CASE_ ).strip() if not number: raise ValueError("No input value was provided" ) lowercase__ = "-" if number.startswith("-" ) else "" lowercase__ = number.lstrip("-" ) if not number.isnumeric(): raise ValueError("Input value is not an integer" ) return f'''{negative}0b{binary_recursive(int(SCREAMING_SNAKE_CASE_ ) )}''' if __name__ == "__main__": from doctest import testmod testmod()
224
1
import os import pytest from transformers.dynamic_module_utils import get_imports UpperCamelCase = ''' import os ''' UpperCamelCase = ''' def foo(): import os return False ''' UpperCamelCase = ''' def foo(): def bar(): if True: import os return False return bar() ''' UpperCamelCase = ''' import os try: import bar except ImportError: raise ValueError() ''' UpperCamelCase = ''' import os def foo(): try: import bar except ImportError: raise ValueError() ''' UpperCamelCase = ''' import os try: import bar except (ImportError, AttributeError): raise ValueError() ''' UpperCamelCase = ''' import os try: import bar except ImportError as e: raise ValueError() ''' UpperCamelCase = ''' import os try: import bar except: raise ValueError() ''' UpperCamelCase = ''' import os try: import bar import baz except ImportError: raise ValueError() ''' UpperCamelCase = ''' import os try: import bar import baz except ImportError: x = 1 raise ValueError() ''' UpperCamelCase = [ TOP_LEVEL_IMPORT, IMPORT_IN_FUNCTION, DEEPLY_NESTED_IMPORT, TOP_LEVEL_TRY_IMPORT, GENERIC_EXCEPT_IMPORT, MULTILINE_TRY_IMPORT, MULTILINE_BOTH_IMPORT, MULTIPLE_EXCEPTS_IMPORT, EXCEPT_AS_IMPORT, TRY_IMPORT_IN_FUNCTION, ] @pytest.mark.parametrize("case" , _lowerCamelCase) def lowercase_ ( _lowerCamelCase : Union[str, Any] , _lowerCamelCase : List[str]): lowercase__ : Dict = os.path.join(_lowerCamelCase , "test_file.py") with open(_lowerCamelCase , "w") as _tmp_file: _tmp_file.write(_lowerCamelCase) lowercase__ : List[Any] = get_imports(_lowerCamelCase) assert parsed_imports == ["os"]
87
import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() def snake_case_ ( lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Dict , lowerCAmelCase_ : Any , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : int ): # Initialise PyTorch model. # If you want to convert a checkpoint that uses absolute position embeddings, make sure to set reset_position_index_per_cell of # TapasConfig to False. # initialize configuration from json file __lowercase : Optional[int] = TapasConfig.from_json_file(lowerCAmelCase_ ) # set absolute/relative position embeddings parameter __lowercase : Optional[Any] = reset_position_index_per_cell # set remaining parameters of TapasConfig as well as the model based on the task if task == "SQA": __lowercase : Union[str, Any] = TapasForQuestionAnswering(config=lowerCAmelCase_ ) elif task == "WTQ": # run_task_main.py hparams __lowercase : List[Any] = 4 __lowercase : Union[str, Any] = True # hparam_utils.py hparams __lowercase : Any = 0.664_694 __lowercase : Tuple = 0.207_951 __lowercase : Dict = 0.121_194 __lowercase : List[str] = True __lowercase : str = True __lowercase : Dict = False __lowercase : Tuple = 0.0_352_513 __lowercase : List[Any] = TapasForQuestionAnswering(config=lowerCAmelCase_ ) elif task == "WIKISQL_SUPERVISED": # run_task_main.py hparams __lowercase : Optional[int] = 4 __lowercase : int = False # hparam_utils.py hparams __lowercase : Tuple = 36.4_519 __lowercase : str = 0.903_421 __lowercase : List[Any] = 222.088 __lowercase : Union[str, Any] = True __lowercase : Tuple = True __lowercase : Union[str, Any] = True __lowercase : Optional[Any] = 0.763_141 __lowercase : str = TapasForQuestionAnswering(config=lowerCAmelCase_ ) elif task == "TABFACT": __lowercase : List[Any] = TapasForSequenceClassification(config=lowerCAmelCase_ ) elif task == "MLM": __lowercase : Optional[int] = TapasForMaskedLM(config=lowerCAmelCase_ ) elif task == "INTERMEDIATE_PRETRAINING": __lowercase : Dict = TapasModel(config=lowerCAmelCase_ ) else: raise ValueError(F"Task {task} not supported." ) print(F"Building PyTorch model from configuration: {config}" ) # Load weights from tf checkpoint load_tf_weights_in_tapas(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # Save pytorch-model (weights and configuration) print(F"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(lowerCAmelCase_ ) # Save tokenizer files print(F"Save tokenizer files to {pytorch_dump_path}" ) __lowercase : Any = TapasTokenizer(vocab_file=tf_checkpoint_path[:-10] + """vocab.txt""" , model_max_length=512 ) tokenizer.save_pretrained(lowerCAmelCase_ ) print("""Used relative position embeddings:""" , model.config.reset_position_index_per_cell ) if __name__ == "__main__": lowerCamelCase : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--task''', default='''SQA''', type=str, help='''Model task for which to convert a checkpoint. Defaults to SQA.''' ) parser.add_argument( '''--reset_position_index_per_cell''', default=False, action='''store_true''', help='''Whether to use relative position embeddings or not. Defaults to True.''', ) parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--tapas_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained TAPAS model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCamelCase : int = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.task, args.reset_position_index_per_cell, args.tf_checkpoint_path, args.tapas_config_file, args.pytorch_dump_path, )
233
0
# Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position lowerCamelCase__ : Any = '''2.13.1''' import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse('''3.7'''): raise ImportWarning( '''To use `datasets`, Python>=3.7 is required, and the current version of Python doesn\'t match this condition.''' ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( '''To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn\'t match this condition.\n''' '''If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.''' ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip lowerCamelCase__ : Dict = concatenate_datasets lowerCamelCase__ : Dict = DownloadConfig lowerCamelCase__ : Dict = DownloadManager lowerCamelCase__ : Dict = DownloadMode lowerCamelCase__ : Optional[int] = DownloadConfig lowerCamelCase__ : Dict = DownloadMode lowerCamelCase__ : Dict = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
350
import math def __lowerCamelCase ( ) -> None: """simple docstring""" A__ = input("""Enter message: """ ) A__ = int(input(F'Enter key [2-{len(__a ) - 1}]: ' ) ) A__ = input("""Encryption/Decryption [e/d]: """ ) if mode.lower().startswith("""e""" ): A__ = encrypt_message(__a , __a ) elif mode.lower().startswith("""d""" ): A__ = decrypt_message(__a , __a ) # Append pipe symbol (vertical bar) to identify spaces at the end. print(F'Output:\n{text + "|"}' ) def __lowerCamelCase ( __a :int , __a :str ) -> str: """simple docstring""" A__ = [""""""] * key for col in range(__a ): A__ = col while pointer < len(__a ): cipher_text[col] += message[pointer] pointer += key return "".join(__a ) def __lowerCamelCase ( __a :int , __a :str ) -> str: """simple docstring""" A__ = math.ceil(len(__a ) / key ) A__ = key A__ = (num_cols * num_rows) - len(__a ) A__ = [""""""] * num_cols A__ = 0 A__ = 0 for symbol in message: plain_text[col] += symbol col += 1 if ( (col == num_cols) or (col == num_cols - 1) and (row >= num_rows - num_shaded_boxes) ): A__ = 0 row += 1 return "".join(__a ) if __name__ == "__main__": import doctest doctest.testmod() main()
276
0
from __future__ import annotations import unittest from transformers import DistilBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.distilbert.modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertModel, ) class snake_case_ : '''simple docstring''' def __init__( self : Optional[Any] , _UpperCamelCase : Union[str, Any] , ) ->Optional[Any]: snake_case_ = parent snake_case_ = 1_3 snake_case_ = 7 snake_case_ = True snake_case_ = True snake_case_ = False snake_case_ = True snake_case_ = 9_9 snake_case_ = 3_2 snake_case_ = 2 snake_case_ = 4 snake_case_ = 3_7 snake_case_ = '''gelu''' snake_case_ = 0.1 snake_case_ = 0.1 snake_case_ = 5_1_2 snake_case_ = 1_6 snake_case_ = 2 snake_case_ = 0.02 snake_case_ = 3 snake_case_ = 4 snake_case_ = None def snake_case__( self : str ) ->str: snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ = None if self.use_input_mask: snake_case_ = random_attention_mask([self.batch_size, self.seq_length] ) snake_case_ = None snake_case_ = None snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case_ = ids_tensor([self.batch_size] , self.num_choices ) snake_case_ = DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def snake_case__( self : Dict , _UpperCamelCase : Optional[Any] , _UpperCamelCase : List[str] , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : Optional[Any] , _UpperCamelCase : Dict , _UpperCamelCase : Dict ) ->int: snake_case_ = TFDistilBertModel(config=_UpperCamelCase ) snake_case_ = {'''input_ids''': input_ids, '''attention_mask''': input_mask} snake_case_ = model(_UpperCamelCase ) snake_case_ = [input_ids, input_mask] snake_case_ = model(_UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__( self : Tuple , _UpperCamelCase : List[str] , _UpperCamelCase : Any , _UpperCamelCase : Dict , _UpperCamelCase : Optional[Any] , _UpperCamelCase : Tuple , _UpperCamelCase : Union[str, Any] ) ->Dict: snake_case_ = TFDistilBertForMaskedLM(config=_UpperCamelCase ) snake_case_ = {'''input_ids''': input_ids, '''attention_mask''': input_mask} snake_case_ = model(_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case__( self : Tuple , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : str , _UpperCamelCase : Optional[Any] , _UpperCamelCase : Optional[Any] , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : Dict ) ->str: snake_case_ = TFDistilBertForQuestionAnswering(config=_UpperCamelCase ) snake_case_ = { '''input_ids''': input_ids, '''attention_mask''': input_mask, } snake_case_ = model(_UpperCamelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def snake_case__( self : Union[str, Any] , _UpperCamelCase : Any , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : str , _UpperCamelCase : Any , _UpperCamelCase : Optional[int] ) ->List[str]: snake_case_ = self.num_labels snake_case_ = TFDistilBertForSequenceClassification(_UpperCamelCase ) snake_case_ = {'''input_ids''': input_ids, '''attention_mask''': input_mask} snake_case_ = model(_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case__( self : Optional[Any] , _UpperCamelCase : Any , _UpperCamelCase : int , _UpperCamelCase : List[Any] , _UpperCamelCase : str , _UpperCamelCase : Any , _UpperCamelCase : str ) ->Optional[Any]: snake_case_ = self.num_choices snake_case_ = TFDistilBertForMultipleChoice(_UpperCamelCase ) snake_case_ = tf.tile(tf.expand_dims(_UpperCamelCase , 1 ) , (1, self.num_choices, 1) ) snake_case_ = tf.tile(tf.expand_dims(_UpperCamelCase , 1 ) , (1, self.num_choices, 1) ) snake_case_ = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, } snake_case_ = model(_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def snake_case__( self : Optional[int] , _UpperCamelCase : Optional[Any] , _UpperCamelCase : Optional[int] , _UpperCamelCase : int , _UpperCamelCase : List[Any] , _UpperCamelCase : int , _UpperCamelCase : List[str] ) ->int: snake_case_ = self.num_labels snake_case_ = TFDistilBertForTokenClassification(_UpperCamelCase ) snake_case_ = {'''input_ids''': input_ids, '''attention_mask''': input_mask} snake_case_ = model(_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def snake_case__( self : int ) ->Optional[Any]: snake_case_ = self.prepare_config_and_inputs() ((snake_case_), (snake_case_), (snake_case_), (snake_case_), (snake_case_), (snake_case_)) = config_and_inputs snake_case_ = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class snake_case_ ( __A , __A , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[Any] = ( ( TFDistilBertModel, TFDistilBertForMaskedLM, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertForMultipleChoice, ) if is_tf_available() else None ) SCREAMING_SNAKE_CASE : List[Any] = ( { 'feature-extraction': TFDistilBertModel, 'fill-mask': TFDistilBertForMaskedLM, 'question-answering': TFDistilBertForQuestionAnswering, 'text-classification': TFDistilBertForSequenceClassification, 'token-classification': TFDistilBertForTokenClassification, 'zero-shot': TFDistilBertForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE : List[Any] = False SCREAMING_SNAKE_CASE : Dict = False def snake_case__( self : int ) ->Tuple: snake_case_ = TFDistilBertModelTester(self ) snake_case_ = ConfigTester(self , config_class=_UpperCamelCase , dim=3_7 ) def snake_case__( self : int ) ->Tuple: self.config_tester.run_common_tests() def snake_case__( self : Any ) ->Tuple: snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*_UpperCamelCase ) def snake_case__( self : Tuple ) ->List[str]: snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*_UpperCamelCase ) def snake_case__( self : List[str] ) ->Optional[Any]: snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*_UpperCamelCase ) def snake_case__( self : int ) ->Optional[Any]: snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*_UpperCamelCase ) def snake_case__( self : Union[str, Any] ) ->List[Any]: snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*_UpperCamelCase ) def snake_case__( self : Tuple ) ->Optional[int]: snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*_UpperCamelCase ) @slow def snake_case__( self : Dict ) ->int: for model_name in list(TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1] ): snake_case_ = TFDistilBertModel.from_pretrained(_UpperCamelCase ) self.assertIsNotNone(_UpperCamelCase ) @require_tf class snake_case_ ( unittest.TestCase ): '''simple docstring''' @slow def snake_case__( self : List[str] ) ->str: snake_case_ = TFDistilBertModel.from_pretrained('''distilbert-base-uncased''' ) snake_case_ = tf.constant([[0, 1, 2, 3, 4, 5]] ) snake_case_ = model(_UpperCamelCase )[0] snake_case_ = [1, 6, 7_6_8] self.assertEqual(output.shape , _UpperCamelCase ) snake_case_ = tf.constant( [ [ [0.19261885, -0.13732955, 0.4119799], [0.22150156, -0.07422661, 0.39037204], [0.22756018, -0.0896414, 0.3701467], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , _UpperCamelCase , atol=1e-4 )
8
'''simple docstring''' import warnings from contextlib import contextmanager from ....processing_utils import ProcessorMixin class __A ( A ): '''simple docstring''' __lowerCamelCase : Optional[Any] = 'MCTCTFeatureExtractor' __lowerCamelCase : Optional[Any] = 'AutoTokenizer' def __init__(self , A , A ) -> Dict: """simple docstring""" super().__init__(A , A ) _a = self.feature_extractor _a = False def __call__(self , *A , **A ) -> Optional[int]: """simple docstring""" if self._in_target_context_manager: return self.current_processor(*A , **A ) if "raw_speech" in kwargs: warnings.warn('''Using `raw_speech` as a keyword argument is deprecated. Use `audio` instead.''' ) _a = kwargs.pop('''raw_speech''' ) else: _a = kwargs.pop('''audio''' , A ) _a = kwargs.pop('''sampling_rate''' , A ) _a = kwargs.pop('''text''' , A ) if len(A ) > 0: _a = args[0] _a = args[1:] if audio is None and text is None: raise ValueError('''You need to specify either an `audio` or `text` input to process.''' ) if audio is not None: _a = self.feature_extractor(A , *A , sampling_rate=A , **A ) if text is not None: _a = self.tokenizer(A , **A ) if text is None: return inputs elif audio is None: return encodings else: _a = encodings['''input_ids'''] return inputs def a__ (self , *A , **A ) -> int: """simple docstring""" return self.tokenizer.batch_decode(*A , **A ) def a__ (self , *A , **A ) -> Tuple: """simple docstring""" if self._in_target_context_manager: return self.current_processor.pad(*A , **A ) _a = kwargs.pop('''input_features''' , A ) _a = kwargs.pop('''labels''' , A ) if len(A ) > 0: _a = args[0] _a = args[1:] if input_features is not None: _a = self.feature_extractor.pad(A , *A , **A ) if labels is not None: _a = self.tokenizer.pad(A , **A ) if labels is None: return input_features elif input_features is None: return labels else: _a = labels['''input_ids'''] return input_features def a__ (self , *A , **A ) -> List[str]: """simple docstring""" return self.tokenizer.decode(*A , **A ) @contextmanager def a__ (self ) -> Dict: """simple docstring""" warnings.warn( '''`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ''' '''labels by using the argument `text` of the regular `__call__` method (either in the same call as ''' '''your audio inputs, or in a separate call.''' ) _a = True _a = self.tokenizer yield _a = self.feature_extractor _a = False
211
0
from __future__ import annotations from collections.abc import Generator import requests from bsa import BeautifulSoup __lowerCamelCase : int = '''https://www.indeed.co.in/jobs?q=mobile+app+development&l=''' def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : str = "mumbai" ) -> Generator[tuple[str, str], None, None]: """simple docstring""" SCREAMING_SNAKE_CASE__ = BeautifulSoup(requests.get(url + location ).content , """html.parser""" ) # This attribute finds out all the specifics listed in a job for job in soup.find_all("""div""" , attrs={"""data-tn-component""": """organicJob"""} ): SCREAMING_SNAKE_CASE__ = job.find("""a""" , attrs={"""data-tn-element""": """jobTitle"""} ).text.strip() SCREAMING_SNAKE_CASE__ = job.find("""span""" , {"""class""": """company"""} ).text.strip() yield job_title, company_name if __name__ == "__main__": for i, job in enumerate(fetch_jobs('''Bangalore'''), 1): print(F"""Job {i:>2} is {job[0]} at {job[1]}""")
204
import os from argparse import ArgumentParser from typing import List import torch.utils.data from datasets import Dataset, IterableDataset from datasets.distributed import split_dataset_by_node __lowerCamelCase : int = 4 __lowerCamelCase : Dict = 3 class __snake_case ( lowerCamelCase_ ): pass def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : List[str] ) -> List[str]: """simple docstring""" for shard in shards: for i in range(__UpperCamelCase ): yield {"i": i, "shard": shard} def __SCREAMING_SNAKE_CASE ( ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = int(os.environ["""RANK"""] ) SCREAMING_SNAKE_CASE__ = int(os.environ["""WORLD_SIZE"""] ) SCREAMING_SNAKE_CASE__ = ArgumentParser() parser.add_argument("""--streaming""" , type=__UpperCamelCase ) parser.add_argument("""--local_rank""" , type=__UpperCamelCase ) parser.add_argument("""--num_workers""" , type=__UpperCamelCase , default=0 ) SCREAMING_SNAKE_CASE__ = parser.parse_args() SCREAMING_SNAKE_CASE__ = args.streaming SCREAMING_SNAKE_CASE__ = args.num_workers SCREAMING_SNAKE_CASE__ = {"""shards""": [f"""shard_{shard_idx}""" for shard_idx in range(__UpperCamelCase )]} SCREAMING_SNAKE_CASE__ = IterableDataset.from_generator(__UpperCamelCase , gen_kwargs=__UpperCamelCase ) if not streaming: SCREAMING_SNAKE_CASE__ = Dataset.from_list(list(__UpperCamelCase ) ) SCREAMING_SNAKE_CASE__ = split_dataset_by_node(__UpperCamelCase , rank=__UpperCamelCase , world_size=__UpperCamelCase ) SCREAMING_SNAKE_CASE__ = torch.utils.data.DataLoader(__UpperCamelCase , num_workers=__UpperCamelCase ) SCREAMING_SNAKE_CASE__ = NUM_SHARDS * NUM_ITEMS_PER_SHARD SCREAMING_SNAKE_CASE__ = full_size // world_size expected_local_size += int(rank < (full_size % world_size) ) SCREAMING_SNAKE_CASE__ = sum(1 for _ in dataloader ) if local_size != expected_local_size: raise FailedTestError(f"""local_size {local_size} != expected_local_size {expected_local_size}""" ) if __name__ == "__main__": main()
204
1
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_barthez import BarthezTokenizer else: A : Any = None A : Dict = logging.get_logger(__name__) A : List[Any] = {'vocab_file': 'sentencepiece.bpe.model', 'tokenizer_file': 'tokenizer.json'} A : List[Any] = { 'vocab_file': { 'moussaKam/mbarthez': 'https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model', 'moussaKam/barthez': 'https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model', 'moussaKam/barthez-orangesum-title': ( 'https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model' ), }, 'tokenizer_file': { 'moussaKam/mbarthez': 'https://huggingface.co/moussaKam/mbarthez/resolve/main/tokenizer.json', 'moussaKam/barthez': 'https://huggingface.co/moussaKam/barthez/resolve/main/tokenizer.json', 'moussaKam/barthez-orangesum-title': ( 'https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/tokenizer.json' ), }, } A : List[str] = { 'moussaKam/mbarthez': 1_0_2_4, 'moussaKam/barthez': 1_0_2_4, 'moussaKam/barthez-orangesum-title': 1_0_2_4, } A : str = '▁' class __A( SCREAMING_SNAKE_CASE_ ): snake_case_ = VOCAB_FILES_NAMES snake_case_ = PRETRAINED_VOCAB_FILES_MAP snake_case_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ = ['''input_ids''', '''attention_mask'''] snake_case_ = BarthezTokenizer def __init__( self , _snake_case=None , _snake_case=None , _snake_case="<s>" , _snake_case="</s>" , _snake_case="</s>" , _snake_case="<s>" , _snake_case="<unk>" , _snake_case="<pad>" , _snake_case="<mask>" , **_snake_case , ) -> Dict: '''simple docstring''' __a = AddedToken(__UpperCAmelCase , lstrip=__UpperCAmelCase , rstrip=__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else mask_token super().__init__( __UpperCAmelCase , tokenizer_file=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , unk_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , **__UpperCAmelCase , ) __a = vocab_file __a = False if not self.vocab_file else True def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case = None ) -> List[int]: '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __a = [self.cls_token_id] __a = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case = None ) -> List[int]: '''simple docstring''' __a = [self.sep_token_id] __a = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case = None ) -> Tuple[str]: '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(__UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return __a = os.path.join( __UpperCAmelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCAmelCase ): copyfile(self.vocab_file , __UpperCAmelCase ) return (out_vocab_file,)
6
'''simple docstring''' import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch _lowerCAmelCase = logging.get_logger(__name__) class lowerCAmelCase_: '''simple docstring''' def __init__( self ,__UpperCAmelCase = None ,__UpperCAmelCase = None ,__UpperCAmelCase=None ,__UpperCAmelCase=None ) -> str: if not conversation_id: lowerCAmelCase__ : List[str] = uuid.uuida() if past_user_inputs is None: lowerCAmelCase__ : List[Any] = [] if generated_responses is None: lowerCAmelCase__ : str = [] lowerCAmelCase__ : uuid.UUID = conversation_id lowerCAmelCase__ : List[str] = past_user_inputs lowerCAmelCase__ : List[str] = generated_responses lowerCAmelCase__ : Optional[str] = text def __eq__( self ,__UpperCAmelCase ) -> Dict: if not isinstance(__UpperCAmelCase ,__UpperCAmelCase ): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase = False ) -> Optional[Any]: if self.new_user_input: if overwrite: logger.warning( F"""User input added while unprocessed input was existing: \"{self.new_user_input}\" was overwritten """ F"""with: \"{text}\".""" ) lowerCAmelCase__ : Optional[int] = text else: logger.warning( F"""User input added while unprocessed input was existing: \"{self.new_user_input}\" new input """ F"""ignored: \"{text}\". Set `overwrite` to True to overwrite unprocessed user input""" ) else: lowerCAmelCase__ : Optional[Any] = text def UpperCAmelCase_ ( self ) -> List[Any]: if self.new_user_input: self.past_user_inputs.append(self.new_user_input ) lowerCAmelCase__ : Union[str, Any] = None def UpperCAmelCase_ ( self ,__UpperCAmelCase ) -> Tuple: self.generated_responses.append(__UpperCAmelCase ) def UpperCAmelCase_ ( self ) -> List[str]: for user_input, generated_response in zip(self.past_user_inputs ,self.generated_responses ): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self ) -> Tuple: lowerCAmelCase__ : Tuple = F"""Conversation id: {self.uuid} \n""" for is_user, text in self.iter_texts(): lowerCAmelCase__ : Any = """user""" if is_user else """bot""" output += F"""{name} >> {text} \n""" return output @add_end_docstrings( SCREAMING_SNAKE_CASE_ , R''' min_length_for_response (`int`, *optional*, defaults to 32): The minimum length (in number of tokens) for a response. minimum_tokens (`int`, *optional*, defaults to 10): The minimum length of tokens to leave for a response. ''' , ) class lowerCAmelCase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' def __init__( self ,*__UpperCAmelCase ,**__UpperCAmelCase ) -> Tuple: super().__init__(*__UpperCAmelCase ,**__UpperCAmelCase ) if self.tokenizer.pad_token_id is None: lowerCAmelCase__ : Tuple = self.tokenizer.eos_token def UpperCAmelCase_ ( self ,__UpperCAmelCase=None ,__UpperCAmelCase=None ,__UpperCAmelCase=None ,**__UpperCAmelCase ) -> Optional[int]: lowerCAmelCase__ : List[Any] = {} lowerCAmelCase__ : Optional[int] = {} lowerCAmelCase__ : List[str] = {} if min_length_for_response is not None: lowerCAmelCase__ : Any = min_length_for_response if minimum_tokens is not None: lowerCAmelCase__ : Optional[int] = minimum_tokens if "max_length" in generate_kwargs: lowerCAmelCase__ : Optional[Any] = generate_kwargs["""max_length"""] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: lowerCAmelCase__ : int = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(__UpperCAmelCase ) return preprocess_params, forward_params, postprocess_params def __call__( self ,__UpperCAmelCase ,__UpperCAmelCase=0 ,**__UpperCAmelCase ) -> List[str]: lowerCAmelCase__ : Optional[int] = super().__call__(__UpperCAmelCase ,num_workers=__UpperCAmelCase ,**__UpperCAmelCase ) if isinstance(__UpperCAmelCase ,__UpperCAmelCase ) and len(__UpperCAmelCase ) == 1: return outputs[0] return outputs def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase=32 ) -> Dict[str, Any]: if not isinstance(__UpperCAmelCase ,__UpperCAmelCase ): raise ValueError("""ConversationalPipeline, expects Conversation as inputs""" ) if conversation.new_user_input is None: raise ValueError( F"""Conversation with UUID {type(conversation.uuid )} does not contain new user input to process. """ """Add user inputs with the conversation's `add_user_input` method""" ) if hasattr(self.tokenizer ,"""_build_conversation_input_ids""" ): lowerCAmelCase__ : str = self.tokenizer._build_conversation_input_ids(__UpperCAmelCase ) else: # If the tokenizer cannot handle conversations, we default to only the old version lowerCAmelCase__ : List[Any] = self._legacy_parse_and_tokenize(__UpperCAmelCase ) if self.framework == "pt": lowerCAmelCase__ : List[Any] = torch.LongTensor([input_ids] ) elif self.framework == "tf": lowerCAmelCase__ : Dict = tf.constant([input_ids] ) return {"input_ids": input_ids, "conversation": conversation} def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase=10 ,**__UpperCAmelCase ) -> Dict: lowerCAmelCase__ : Optional[Any] = generate_kwargs.get("""max_length""" ,self.model.config.max_length ) lowerCAmelCase__ : Optional[Any] = model_inputs["""input_ids"""].shape[1] if max_length - minimum_tokens < n: logger.warning(F"""Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})""" ) lowerCAmelCase__ : str = max_length - minimum_tokens lowerCAmelCase__ : Union[str, Any] = model_inputs["""input_ids"""][:, -trim:] if "attention_mask" in model_inputs: lowerCAmelCase__ : Tuple = model_inputs["""attention_mask"""][:, -trim:] lowerCAmelCase__ : str = model_inputs.pop("""conversation""" ) lowerCAmelCase__ : Union[str, Any] = max_length lowerCAmelCase__ : Any = self.model.generate(**__UpperCAmelCase ,**__UpperCAmelCase ) if self.model.config.is_encoder_decoder: lowerCAmelCase__ : int = 1 else: lowerCAmelCase__ : int = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase=True ) -> List[str]: lowerCAmelCase__ : Optional[int] = model_outputs["""output_ids"""] lowerCAmelCase__ : Tuple = self.tokenizer.decode( output_ids[0] ,skip_special_tokens=__UpperCAmelCase ,clean_up_tokenization_spaces=__UpperCAmelCase ,) lowerCAmelCase__ : Union[str, Any] = model_outputs["""conversation"""] conversation.mark_processed() conversation.append_response(__UpperCAmelCase ) return conversation def UpperCAmelCase_ ( self ,__UpperCAmelCase ) -> Dict: lowerCAmelCase__ : Dict = self.tokenizer.eos_token_id lowerCAmelCase__ : int = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(__UpperCAmelCase ,add_special_tokens=__UpperCAmelCase ) + [eos_token_id] ) else: input_ids.extend(self.tokenizer.encode(__UpperCAmelCase ,add_special_tokens=__UpperCAmelCase ) ) if len(__UpperCAmelCase ) > self.tokenizer.model_max_length: lowerCAmelCase__ : Optional[Any] = input_ids[-self.tokenizer.model_max_length :] return input_ids
37
0
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin _A = random.Random() def UpperCAmelCase ( a_, a_=1.0, a_=None, a_=None ): '''simple docstring''' if rng is None: lowerCamelCase : Tuple = global_rng lowerCamelCase : Any = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class _lowercase ( unittest.TestCase ): def __init__( self , UpperCAmelCase_ , UpperCAmelCase_=7 , UpperCAmelCase_=400 , UpperCAmelCase_=2000 , UpperCAmelCase_=1 , UpperCAmelCase_=0.0 , UpperCAmelCase_=16000 , UpperCAmelCase_=True , UpperCAmelCase_=True , ) -> List[str]: lowerCamelCase : Dict = parent lowerCamelCase : Any = batch_size lowerCamelCase : Tuple = min_seq_length lowerCamelCase : Tuple = max_seq_length lowerCamelCase : int = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) lowerCamelCase : Union[str, Any] = feature_size lowerCamelCase : Optional[int] = padding_value lowerCamelCase : str = sampling_rate lowerCamelCase : Union[str, Any] = return_attention_mask lowerCamelCase : Union[str, Any] = do_normalize def _UpperCamelCase ( self ) -> List[str]: return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def _UpperCamelCase ( self , UpperCAmelCase_=False , UpperCAmelCase_=False ) -> Optional[Any]: def _flatten(UpperCAmelCase_ ): return list(itertools.chain(*UpperCAmelCase_ ) ) if equal_length: lowerCamelCase : Optional[Any] = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size lowerCamelCase : List[Any] = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: lowerCamelCase : Optional[Any] = [np.asarray(UpperCAmelCase_ ) for x in speech_inputs] return speech_inputs class _lowercase ( __UpperCAmelCase , unittest.TestCase ): lowercase_ = WavaVecaFeatureExtractor def _UpperCamelCase ( self ) -> Optional[Any]: lowerCamelCase : Optional[int] = WavaVecaFeatureExtractionTester(self ) def _UpperCamelCase ( self , UpperCAmelCase_ ) -> Dict: self.assertTrue(np.all(np.mean(UpperCAmelCase_ , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(UpperCAmelCase_ , axis=0 ) - 1 ) < 1E-3 ) ) def _UpperCamelCase ( self ) -> List[Any]: # Tests that all call wrap to encode_plus and batch_encode_plus lowerCamelCase : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 lowerCamelCase : List[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase : str = [np.asarray(UpperCAmelCase_ ) for speech_input in speech_inputs] # Test not batched input lowerCamelCase : List[str] = feat_extract(speech_inputs[0] , return_tensors='np' ).input_values lowerCamelCase : Optional[int] = feat_extract(np_speech_inputs[0] , return_tensors='np' ).input_values self.assertTrue(np.allclose(UpperCAmelCase_ , UpperCAmelCase_ , atol=1E-3 ) ) # Test batched lowerCamelCase : str = feat_extract(UpperCAmelCase_ , return_tensors='np' ).input_values lowerCamelCase : Any = feat_extract(UpperCAmelCase_ , return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(UpperCAmelCase_ , UpperCAmelCase_ ): self.assertTrue(np.allclose(UpperCAmelCase_ , UpperCAmelCase_ , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. lowerCamelCase : Union[str, Any] = [floats_list((1, x) )[0] for x in (800, 800, 800)] lowerCamelCase : Optional[int] = np.asarray(UpperCAmelCase_ ) lowerCamelCase : Dict = feat_extract(UpperCAmelCase_ , return_tensors='np' ).input_values lowerCamelCase : Tuple = feat_extract(UpperCAmelCase_ , return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(UpperCAmelCase_ , UpperCAmelCase_ ): self.assertTrue(np.allclose(UpperCAmelCase_ , UpperCAmelCase_ , atol=1E-3 ) ) def _UpperCamelCase ( self ) -> str: lowerCamelCase : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase : Any = ['longest', 'max_length', 'do_not_pad'] lowerCamelCase : Tuple = [None, 1600, None] for max_length, padding in zip(UpperCAmelCase_ , UpperCAmelCase_ ): lowerCamelCase : Dict = feat_extract(UpperCAmelCase_ , padding=UpperCAmelCase_ , max_length=UpperCAmelCase_ , return_tensors='np' ) lowerCamelCase : Union[str, Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def _UpperCamelCase ( self ) -> Optional[int]: lowerCamelCase : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase : int = range(800 , 1400 , 200 ) lowerCamelCase : Tuple = [floats_list((1, x) )[0] for x in lengths] lowerCamelCase : List[str] = ['longest', 'max_length', 'do_not_pad'] lowerCamelCase : Dict = [None, 1600, None] for max_length, padding in zip(UpperCAmelCase_ , UpperCAmelCase_ ): lowerCamelCase : Optional[Any] = feat_extract(UpperCAmelCase_ , max_length=UpperCAmelCase_ , padding=UpperCAmelCase_ ) lowerCamelCase : Any = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def _UpperCamelCase ( self ) -> Optional[int]: lowerCamelCase : Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase : Any = feat_extract( UpperCAmelCase_ , truncation=UpperCAmelCase_ , max_length=1000 , padding='max_length' , return_tensors='np' ) lowerCamelCase : Optional[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def _UpperCamelCase ( self ) -> List[str]: lowerCamelCase : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase : List[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase : Any = feat_extract( UpperCAmelCase_ , truncation=UpperCAmelCase_ , max_length=1000 , padding='longest' , return_tensors='np' ) lowerCamelCase : Dict = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) lowerCamelCase : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCamelCase : Tuple = feat_extract( UpperCAmelCase_ , truncation=UpperCAmelCase_ , max_length=2000 , padding='longest' , return_tensors='np' ) lowerCamelCase : int = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) @require_torch def _UpperCamelCase ( self ) -> Any: import torch lowerCamelCase : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCamelCase : Optional[int] = np.random.rand(100 ).astype(np.floataa ) lowerCamelCase : Tuple = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: lowerCamelCase : List[Any] = feature_extractor.pad([{'input_values': inputs}] , return_tensors='np' ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) lowerCamelCase : List[str] = feature_extractor.pad([{'input_values': inputs}] , return_tensors='pt' ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) @slow @require_torch def _UpperCamelCase ( self ) -> Optional[Any]: # this test makes sure that models that are using # group norm don't have their feature extractor return the # attention_mask for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: lowerCamelCase : int = WavaVecaConfig.from_pretrained(UpperCAmelCase_ ) lowerCamelCase : Union[str, Any] = WavaVecaFeatureExtractor.from_pretrained(UpperCAmelCase_ ) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask , config.feat_extract_norm == 'layer' )
352
"""simple docstring""" import numpy as np def UpperCAmelCase ( a_, a_, a_ = 1E-12, a_ = 100, ): '''simple docstring''' assert np.shape(a_ )[0] == np.shape(a_ )[1] # Ensure proper dimensionality. assert np.shape(a_ )[0] == np.shape(a_ )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(a_ ) == np.iscomplexobj(a_ ) lowerCamelCase : Optional[int] = np.iscomplexobj(a_ ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(a_, input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. lowerCamelCase : Union[str, Any] = False lowerCamelCase : List[str] = 0 lowerCamelCase : Any = 0 lowerCamelCase : Dict = 1E12 while not convergence: # Multiple matrix by the vector. lowerCamelCase : Optional[int] = np.dot(a_, a_ ) # Normalize the resulting output vector. lowerCamelCase : Optional[int] = w / np.linalg.norm(a_ ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) lowerCamelCase : Optional[Any] = vector.conj().T if is_complex else vector.T lowerCamelCase : str = np.dot(a_, np.dot(a_, a_ ) ) # Check convergence. lowerCamelCase : Optional[int] = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: lowerCamelCase : int = True lowerCamelCase : Optional[Any] = lambda_ if is_complex: lowerCamelCase : Any = np.real(lambda_ ) return lambda_, vector def UpperCAmelCase ( ): '''simple docstring''' lowerCamelCase : str = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) lowerCamelCase : str = np.array([41, 4, 20] ) lowerCamelCase : Optional[Any] = real_input_matrix.astype(np.complexaaa ) lowerCamelCase : Dict = np.triu(1j * complex_input_matrix, 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T lowerCamelCase : List[Any] = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": lowerCamelCase : str = real_input_matrix lowerCamelCase : Any = real_vector elif problem_type == "complex": lowerCamelCase : str = complex_input_matrix lowerCamelCase : Dict = complex_vector # Our implementation. lowerCamelCase , lowerCamelCase : List[str] = power_iteration(a_, a_ ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). lowerCamelCase , lowerCamelCase : Optional[Any] = np.linalg.eigh(a_ ) # Last eigenvalue is the maximum one. lowerCamelCase : Dict = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. lowerCamelCase : List[str] = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1E-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(a_ ) - np.abs(a_ ) ) <= 1E-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
205
0
'''simple docstring''' def __A ( lowerCamelCase_ ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = abs(_UpperCAmelCase ) SCREAMING_SNAKE_CASE : List[Any] = 0 while n > 0: res += n % 10 n //= 10 return res def __A ( lowerCamelCase_ ): """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = abs(_UpperCAmelCase ) return n if n < 10 else n % 10 + sum_of_digits(n // 10 ) def __A ( lowerCamelCase_ ): """simple docstring""" return sum(int(_UpperCAmelCase ) for c in str(abs(_UpperCAmelCase ) ) ) def __A ( ): """simple docstring""" from collections.abc import Callable from timeit import timeit def benchmark_a_function(lowerCamelCase_ , lowerCamelCase_ ) -> None: SCREAMING_SNAKE_CASE : str = f'''{func.__name__}({value})''' SCREAMING_SNAKE_CASE : Optional[int] = timeit(f'''__main__.{call}''' , setup="""import __main__""" ) print(f'''{call:56} = {func(_UpperCAmelCase )} -- {timing:.4f} seconds''' ) for value in (26_21_44, 11_25_89_99_06_84_26_24, 1_26_76_50_60_02_28_22_94_01_49_67_03_20_53_76): for func in (sum_of_digits, sum_of_digits_recursion, sum_of_digits_compact): benchmark_a_function(_UpperCAmelCase , _UpperCAmelCase ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
323
import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEIGHTS_NAME, cached_file, get_file_from_repo, has_file, ) a_ : Dict = 'hf-internal-testing/tiny-random-bert' a_ : Tuple = os.path.join(TRANSFORMERS_CACHE, 'models--hf-internal-testing--tiny-random-bert') a_ : Optional[int] = '9b8c223d42b2188cb49d29af482996f9d0f3e5a6' class _snake_case ( unittest.TestCase ): def SCREAMING_SNAKE_CASE__ ( self) -> List[Any]: SCREAMING_SNAKE_CASE = cached_file(a , a) # Should have downloaded the file in here self.assertTrue(os.path.isdir(a)) # Cache should contain at least those three subfolders: for subfolder in ["blobs", "refs", "snapshots"]: self.assertTrue(os.path.isdir(os.path.join(a , a))) with open(os.path.join(a , 'refs' , 'main')) as f: SCREAMING_SNAKE_CASE = f.read() self.assertEqual(a , os.path.join(a , 'snapshots' , a , a)) self.assertTrue(os.path.isfile(a)) # File is cached at the same place the second time. SCREAMING_SNAKE_CASE = cached_file(a , a) self.assertEqual(a , a) # Using a specific revision to test the full commit hash. SCREAMING_SNAKE_CASE = cached_file(a , a , revision='9b8c223') self.assertEqual(a , os.path.join(a , 'snapshots' , a , a)) def SCREAMING_SNAKE_CASE__ ( self) -> List[str]: with self.assertRaisesRegex(a , 'is not a valid model identifier'): SCREAMING_SNAKE_CASE = cached_file('tiny-random-bert' , a) with self.assertRaisesRegex(a , 'is not a valid git identifier'): SCREAMING_SNAKE_CASE = cached_file(a , a , revision='aaaa') with self.assertRaisesRegex(a , 'does not appear to have a file named'): SCREAMING_SNAKE_CASE = cached_file(a , 'conf') def SCREAMING_SNAKE_CASE__ ( self) -> List[Any]: with self.assertRaisesRegex(a , 'does not appear to have a file named'): SCREAMING_SNAKE_CASE = cached_file(a , 'conf') with open(os.path.join(a , 'refs' , 'main')) as f: SCREAMING_SNAKE_CASE = f.read() self.assertTrue(os.path.isfile(os.path.join(a , '.no_exist' , a , 'conf'))) SCREAMING_SNAKE_CASE = cached_file(a , 'conf' , _raise_exceptions_for_missing_entries=a) self.assertIsNone(a) SCREAMING_SNAKE_CASE = cached_file(a , 'conf' , local_files_only=a , _raise_exceptions_for_missing_entries=a) self.assertIsNone(a) SCREAMING_SNAKE_CASE = mock.Mock() SCREAMING_SNAKE_CASE = 500 SCREAMING_SNAKE_CASE = {} SCREAMING_SNAKE_CASE = HTTPError SCREAMING_SNAKE_CASE = {} # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('requests.Session.request' , return_value=a) as mock_head: SCREAMING_SNAKE_CASE = cached_file(a , 'conf' , _raise_exceptions_for_connection_errors=a) self.assertIsNone(a) # This check we did call the fake head request mock_head.assert_called() def SCREAMING_SNAKE_CASE__ ( self) -> int: self.assertTrue(has_file('hf-internal-testing/tiny-bert-pt-only' , a)) self.assertFalse(has_file('hf-internal-testing/tiny-bert-pt-only' , a)) self.assertFalse(has_file('hf-internal-testing/tiny-bert-pt-only' , a)) def SCREAMING_SNAKE_CASE__ ( self) -> Any: # `get_file_from_repo` returns None if the file does not exist self.assertIsNone(get_file_from_repo('bert-base-cased' , 'ahah.txt')) # The function raises if the repository does not exist. with self.assertRaisesRegex(a , 'is not a valid model identifier'): get_file_from_repo('bert-base-case' , a) # The function raises if the revision does not exist. with self.assertRaisesRegex(a , 'is not a valid git identifier'): get_file_from_repo('bert-base-cased' , a , revision='ahaha') SCREAMING_SNAKE_CASE = get_file_from_repo('bert-base-cased' , a) # The name is the cached name which is not very easy to test, so instead we load the content. SCREAMING_SNAKE_CASE = json.loads(open(a , 'r').read()) self.assertEqual(config['hidden_size'] , 768) def SCREAMING_SNAKE_CASE__ ( self) -> Optional[int]: with tempfile.TemporaryDirectory() as tmp_dir: SCREAMING_SNAKE_CASE = Path(a) / 'a.txt' filename.touch() self.assertEqual(get_file_from_repo(a , 'a.txt') , str(a)) self.assertIsNone(get_file_from_repo(a , 'b.txt'))
137
0
"""simple docstring""" from . import ( albert, align, altclip, audio_spectrogram_transformer, auto, autoformer, bark, bart, barthez, bartpho, beit, bert, bert_generation, bert_japanese, bertweet, big_bird, bigbird_pegasus, biogpt, bit, blenderbot, blenderbot_small, blip, blip_a, bloom, bridgetower, byta, camembert, canine, chinese_clip, clap, clip, clipseg, codegen, conditional_detr, convbert, convnext, convnextva, cpm, cpmant, ctrl, cvt, dataavec, deberta, deberta_va, decision_transformer, deformable_detr, deit, deprecated, deta, detr, dialogpt, dinat, distilbert, dit, donut, dpr, dpt, efficientformer, efficientnet, electra, encodec, encoder_decoder, ernie, ernie_m, esm, falcon, flaubert, flava, fnet, focalnet, fsmt, funnel, git, glpn, gpta, gpt_bigcode, gpt_neo, gpt_neox, gpt_neox_japanese, gpt_swa, gptj, gptsan_japanese, graphormer, groupvit, herbert, hubert, ibert, imagegpt, informer, instructblip, jukebox, layoutlm, layoutlmva, layoutlmva, layoutxlm, led, levit, lilt, llama, longformer, longta, luke, lxmert, mam_aaa, marian, markuplm, maskaformer, maskformer, mbart, mbartaa, mega, megatron_bert, megatron_gpta, mgp_str, mluke, mobilebert, mobilenet_va, mobilenet_va, mobilevit, mobilevitva, mpnet, mra, mta, musicgen, mvp, nat, nezha, nllb, nllb_moe, nystromformer, oneformer, open_llama, openai, opt, owlvit, pegasus, pegasus_x, perceiver, phobert, pixastruct, plbart, poolformer, prophetnet, qdqbert, rag, realm, reformer, regnet, rembert, resnet, roberta, roberta_prelayernorm, roc_bert, roformer, rwkv, sam, segformer, sew, sew_d, speech_encoder_decoder, speech_to_text, speech_to_text_a, speechta, splinter, squeezebert, swiftformer, swin, swinasr, swinva, switch_transformers, ta, table_transformer, tapas, time_series_transformer, timesformer, timm_backbone, transfo_xl, trocr, tvlt, umta, unispeech, unispeech_sat, upernet, videomae, vilt, vision_encoder_decoder, vision_text_dual_encoder, visual_bert, vit, vit_hybrid, vit_mae, vit_msn, vivit, wavaveca, wavaveca_conformer, wavaveca_phoneme, wavaveca_with_lm, wavlm, whisper, x_clip, xglm, xlm, xlm_prophetnet, xlm_roberta, xlm_roberta_xl, xlnet, xmod, yolos, yoso, )
360
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) SCREAMING_SNAKE_CASE : Any = { '''transfo-xl-wt103''': '''https://huggingface.co/transfo-xl-wt103/resolve/main/config.json''', } class __lowerCamelCase ( __lowercase ): __UpperCamelCase = 'transfo-xl' __UpperCamelCase = ['mems'] __UpperCamelCase = { 'n_token': 'vocab_size', 'hidden_size': 'd_model', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__(self , lowerCamelCase=267_735 , lowerCamelCase=[20_000, 40_000, 200_000] , lowerCamelCase=1_024 , lowerCamelCase=1_024 , lowerCamelCase=16 , lowerCamelCase=64 , lowerCamelCase=4_096 , lowerCamelCase=4 , lowerCamelCase=False , lowerCamelCase=18 , lowerCamelCase=1_600 , lowerCamelCase=1_000 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=0 , lowerCamelCase=-1 , lowerCamelCase=True , lowerCamelCase=0.1 , lowerCamelCase=0.0 , lowerCamelCase=True , lowerCamelCase="normal" , lowerCamelCase=0.01 , lowerCamelCase=0.01 , lowerCamelCase=0.02 , lowerCamelCase=1e-5 , lowerCamelCase=0 , **lowerCamelCase , ): '''simple docstring''' _lowerCAmelCase = vocab_size _lowerCAmelCase = [] self.cutoffs.extend(lowerCamelCase ) if proj_share_all_but_first: _lowerCAmelCase = [False] + [True] * len(self.cutoffs ) else: _lowerCAmelCase = [False] + [False] * len(self.cutoffs ) _lowerCAmelCase = d_model _lowerCAmelCase = d_embed _lowerCAmelCase = d_head _lowerCAmelCase = d_inner _lowerCAmelCase = div_val _lowerCAmelCase = pre_lnorm _lowerCAmelCase = n_layer _lowerCAmelCase = n_head _lowerCAmelCase = mem_len _lowerCAmelCase = same_length _lowerCAmelCase = attn_type _lowerCAmelCase = clamp_len _lowerCAmelCase = sample_softmax _lowerCAmelCase = adaptive _lowerCAmelCase = dropout _lowerCAmelCase = dropatt _lowerCAmelCase = untie_r _lowerCAmelCase = init _lowerCAmelCase = init_range _lowerCAmelCase = proj_init_std _lowerCAmelCase = init_std _lowerCAmelCase = layer_norm_epsilon super().__init__(eos_token_id=lowerCamelCase , **lowerCamelCase ) @property def A__ (self ): '''simple docstring''' logger.info(f"""The model {self.model_type} is one of the few models that has no sequence length limit.""" ) return -1 @max_position_embeddings.setter def A__ (self , lowerCamelCase ): '''simple docstring''' raise NotImplementedError( f"""The model {self.model_type} is one of the few models that has no sequence length limit.""" )
317
0
"""simple docstring""" import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] ) ->str: '''simple docstring''' a, a : str = image.size a, a : Tuple = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 a : Union[str, Any] = image.resize((w, h) , resample=PIL_INTERPOLATION["lanczos"] ) a : int = np.array(_lowercase ).astype(np.floataa ) / 255.0 a : List[str] = image[None].transpose(0 , 3 , 1 , 2 ) a : Dict = torch.from_numpy(_lowercase ) return 2.0 * image - 1.0 class __UpperCamelCase ( a__ ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ) -> str: super().__init__() self.register_modules(vqvae=lowerCAmelCase__ , unet=lowerCAmelCase__ , scheduler=lowerCAmelCase__ ) @torch.no_grad() def __call__( self , lowerCAmelCase__ = None , lowerCAmelCase__ = 1 , lowerCAmelCase__ = 100 , lowerCAmelCase__ = 0.0 , lowerCAmelCase__ = None , lowerCAmelCase__ = "pil" , lowerCAmelCase__ = True , ) -> Union[Tuple, ImagePipelineOutput]: if isinstance(lowerCAmelCase__ , PIL.Image.Image ): a : int = 1 elif isinstance(lowerCAmelCase__ , torch.Tensor ): a : str = image.shape[0] else: raise ValueError(f"""`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(lowerCAmelCase__ )}""" ) if isinstance(lowerCAmelCase__ , PIL.Image.Image ): a : Tuple = preprocess(lowerCAmelCase__ ) a, a : Optional[Any] = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image a : Tuple = (batch_size, self.unet.config.in_channels // 2, height, width) a : List[str] = next(self.unet.parameters() ).dtype a : Any = randn_tensor(lowerCAmelCase__ , generator=lowerCAmelCase__ , device=self.device , dtype=lowerCAmelCase__ ) a : Union[str, Any] = image.to(device=self.device , dtype=lowerCAmelCase__ ) # set timesteps and move to the correct device self.scheduler.set_timesteps(lowerCAmelCase__ , device=self.device ) a : int = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler a : str = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] a : str = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) a : List[Any] = {} if accepts_eta: a : Any = eta for t in self.progress_bar(lowerCAmelCase__ ): # concat latents and low resolution image in the channel dimension. a : str = torch.cat([latents, image] , dim=1 ) a : int = self.scheduler.scale_model_input(lowerCAmelCase__ , lowerCAmelCase__ ) # predict the noise residual a : Union[str, Any] = self.unet(lowerCAmelCase__ , lowerCAmelCase__ ).sample # compute the previous noisy sample x_t -> x_t-1 a : Tuple = self.scheduler.step(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ).prev_sample # decode the image latents with the VQVAE a : List[Any] = self.vqvae.decode(lowerCAmelCase__ ).sample a : int = torch.clamp(lowerCAmelCase__ , -1.0 , 1.0 ) a : Optional[int] = image / 2 + 0.5 a : Optional[int] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": a : Union[str, Any] = self.numpy_to_pil(lowerCAmelCase__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCAmelCase__ )
105
import os import shutil import tempfile import unittest import numpy as np from transformers import AutoTokenizer, BarkProcessor from transformers.testing_utils import require_torch, slow @require_torch class _lowercase ( unittest.TestCase ): '''simple docstring''' def __magic_name__( self :Union[str, Any] ) -> Tuple: __SCREAMING_SNAKE_CASE : str = '''ylacombe/bark-small''' __SCREAMING_SNAKE_CASE : Optional[int] = tempfile.mkdtemp() __SCREAMING_SNAKE_CASE : str = '''en_speaker_1''' __SCREAMING_SNAKE_CASE : Any = '''This is a test string''' __SCREAMING_SNAKE_CASE : int = '''speaker_embeddings_path.json''' __SCREAMING_SNAKE_CASE : int = '''speaker_embeddings''' def __magic_name__( self :List[str] , **lowerCAmelCase__ :Union[str, Any] ) -> Any: return AutoTokenizer.from_pretrained(self.checkpoint , **lowerCAmelCase__ ) def __magic_name__( self :List[str] ) -> int: shutil.rmtree(self.tmpdirname ) def __magic_name__( self :Dict ) -> str: __SCREAMING_SNAKE_CASE : Dict = self.get_tokenizer() __SCREAMING_SNAKE_CASE : Tuple = BarkProcessor(tokenizer=lowerCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) __SCREAMING_SNAKE_CASE : Optional[Any] = BarkProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) @slow def __magic_name__( self :Tuple ) -> List[Any]: __SCREAMING_SNAKE_CASE : Dict = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) processor.save_pretrained( self.tmpdirname , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , speaker_embeddings_directory=self.speaker_embeddings_directory , ) __SCREAMING_SNAKE_CASE : Dict = self.get_tokenizer(bos_token='''(BOS)''' , eos_token='''(EOS)''' ) __SCREAMING_SNAKE_CASE : Union[str, Any] = BarkProcessor.from_pretrained( self.tmpdirname , self.speaker_embeddings_dict_path , bos_token='''(BOS)''' , eos_token='''(EOS)''' , ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) def __magic_name__( self :List[str] ) -> Tuple: __SCREAMING_SNAKE_CASE : List[Any] = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) __SCREAMING_SNAKE_CASE : str = 35 __SCREAMING_SNAKE_CASE : str = 2 __SCREAMING_SNAKE_CASE : List[Any] = 8 __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''semantic_prompt''': np.ones(lowerCAmelCase__ ), '''coarse_prompt''': np.ones((nb_codebooks_coarse, seq_len) ), '''fine_prompt''': np.ones((nb_codebooks_total, seq_len) ), } # test providing already loaded voice_preset __SCREAMING_SNAKE_CASE : Union[str, Any] = processor(text=self.input_string , voice_preset=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : List[str] = inputs['''history_prompt'''] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowerCAmelCase__ , np.array([] ) ).tolist() ) # test loading voice preset from npz file __SCREAMING_SNAKE_CASE : str = os.path.join(self.tmpdirname , '''file.npz''' ) np.savez(lowerCAmelCase__ , **lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : List[Any] = processor(text=self.input_string , voice_preset=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Union[str, Any] = inputs['''history_prompt'''] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowerCAmelCase__ , np.array([] ) ).tolist() ) # test loading voice preset from the hub __SCREAMING_SNAKE_CASE : Union[str, Any] = processor(text=self.input_string , voice_preset=self.voice_preset ) def __magic_name__( self :Tuple ) -> Optional[Any]: __SCREAMING_SNAKE_CASE : Tuple = self.get_tokenizer() __SCREAMING_SNAKE_CASE : Any = BarkProcessor(tokenizer=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : List[Any] = processor(text=self.input_string ) __SCREAMING_SNAKE_CASE : List[Any] = tokenizer( self.input_string , padding='''max_length''' , max_length=256 , add_special_tokens=lowerCAmelCase__ , return_attention_mask=lowerCAmelCase__ , return_token_type_ids=lowerCAmelCase__ , ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key].squeeze().tolist() )
9
0
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCAmelCase_ ( lowerCamelCase_ ): '''simple docstring''' lowerCAmelCase_ : Tuple = ["""image_processor""", """tokenizer"""] lowerCAmelCase_ : Optional[int] = """CLIPImageProcessor""" lowerCAmelCase_ : List[str] = ("""CLIPTokenizer""", """CLIPTokenizerFast""") def __init__( self : int , _UpperCAmelCase : Tuple=None , _UpperCAmelCase : List[Any]=None , **_UpperCAmelCase : Optional[int] ): """simple docstring""" UpperCAmelCase__ = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" , _UpperCAmelCase , ) UpperCAmelCase__ = kwargs.pop("""feature_extractor""" ) UpperCAmelCase__ = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(_UpperCAmelCase , _UpperCAmelCase ) def __call__( self : Any , _UpperCAmelCase : List[Any]=None , _UpperCAmelCase : List[Any]=None , _UpperCAmelCase : Tuple=None , **_UpperCAmelCase : Tuple ): """simple docstring""" if text is None and images is None: raise ValueError("""You have to specify either text or images. Both cannot be none.""" ) if text is not None: UpperCAmelCase__ = self.tokenizer(_UpperCAmelCase , return_tensors=_UpperCAmelCase , **_UpperCAmelCase ) if images is not None: UpperCAmelCase__ = self.image_processor(_UpperCAmelCase , return_tensors=_UpperCAmelCase , **_UpperCAmelCase ) if text is not None and images is not None: UpperCAmelCase__ = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_UpperCAmelCase ) , tensor_type=_UpperCAmelCase ) def SCREAMING_SNAKE_CASE__ ( self : List[str] , *_UpperCAmelCase : int , **_UpperCAmelCase : Optional[Any] ): """simple docstring""" return self.tokenizer.batch_decode(*_UpperCAmelCase , **_UpperCAmelCase ) def SCREAMING_SNAKE_CASE__ ( self : str , *_UpperCAmelCase : List[str] , **_UpperCAmelCase : List[str] ): """simple docstring""" return self.tokenizer.decode(*_UpperCAmelCase , **_UpperCAmelCase ) @property def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ): """simple docstring""" UpperCAmelCase__ = self.tokenizer.model_input_names UpperCAmelCase__ = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def SCREAMING_SNAKE_CASE__ ( self : Tuple ): """simple docstring""" warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""" , _UpperCAmelCase , ) return self.image_processor_class @property def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): """simple docstring""" warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""" , _UpperCAmelCase , ) return self.image_processor
369
'''simple docstring''' def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int ): '''simple docstring''' while b: UpperCAmelCase__ , UpperCAmelCase__ = b, a % b return a def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int ): '''simple docstring''' return a if b == 0 else euclidean_gcd_recursive(SCREAMING_SNAKE_CASE__ , a % b ) def _UpperCamelCase ( ): '''simple docstring''' print(F'''euclidean_gcd(3, 5) = {euclidean_gcd(3 , 5 )}''' ) print(F'''euclidean_gcd(5, 3) = {euclidean_gcd(5 , 3 )}''' ) print(F'''euclidean_gcd(1, 3) = {euclidean_gcd(1 , 3 )}''' ) print(F'''euclidean_gcd(3, 6) = {euclidean_gcd(3 , 6 )}''' ) print(F'''euclidean_gcd(6, 3) = {euclidean_gcd(6 , 3 )}''' ) print(F'''euclidean_gcd_recursive(3, 5) = {euclidean_gcd_recursive(3 , 5 )}''' ) print(F'''euclidean_gcd_recursive(5, 3) = {euclidean_gcd_recursive(5 , 3 )}''' ) print(F'''euclidean_gcd_recursive(1, 3) = {euclidean_gcd_recursive(1 , 3 )}''' ) print(F'''euclidean_gcd_recursive(3, 6) = {euclidean_gcd_recursive(3 , 6 )}''' ) print(F'''euclidean_gcd_recursive(6, 3) = {euclidean_gcd_recursive(6 , 3 )}''' ) if __name__ == "__main__": main()
61
0
import unittest import numpy as np import torch from diffusers import DDIMPipeline, DDIMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow, torch_device from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class __A ( a , unittest.TestCase ): """simple docstring""" UpperCamelCase__ : str =DDIMPipeline UpperCamelCase__ : List[Any] =UNCONDITIONAL_IMAGE_GENERATION_PARAMS UpperCamelCase__ : Tuple =PipelineTesterMixin.required_optional_params - { """num_images_per_prompt""", """latents""", """callback""", """callback_steps""", } UpperCamelCase__ : Tuple =UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS UpperCamelCase__ : Any =False def __lowercase ( self ): """simple docstring""" torch.manual_seed(0 ) __UpperCamelCase : Optional[int] =UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('DownBlock2D', 'AttnDownBlock2D') , up_block_types=('AttnUpBlock2D', 'UpBlock2D') , ) __UpperCamelCase : int =DDIMScheduler() __UpperCamelCase : Optional[int] ={'unet': unet, 'scheduler': scheduler} return components def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=0 ): """simple docstring""" if str(lowerCamelCase__ ).startswith('mps' ): __UpperCamelCase : str =torch.manual_seed(lowerCamelCase__ ) else: __UpperCamelCase : Optional[int] =torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) __UpperCamelCase : Tuple ={ 'batch_size': 1, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Any ='cpu' __UpperCamelCase : Optional[Any] =self.get_dummy_components() __UpperCamelCase : Tuple =self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =self.get_dummy_inputs(lowerCamelCase__ ) __UpperCamelCase : int =pipe(**lowerCamelCase__ ).images __UpperCamelCase : Dict =image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 32, 32, 3) ) __UpperCamelCase : Tuple =np.array( [1.000E00, 5.717E-01, 4.717E-01, 1.000E00, 0.000E00, 1.000E00, 3.000E-04, 0.000E00, 9.000E-04] ) __UpperCamelCase : Tuple =np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCamelCase__ , 1E-3 ) def __lowercase ( self ): """simple docstring""" super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) def __lowercase ( self ): """simple docstring""" super().test_save_load_local(expected_max_difference=3E-3 ) def __lowercase ( self ): """simple docstring""" super().test_save_load_optional_components(expected_max_difference=3E-3 ) def __lowercase ( self ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class __A ( unittest.TestCase ): """simple docstring""" def __lowercase ( self ): """simple docstring""" __UpperCamelCase : str ='google/ddpm-cifar10-32' __UpperCamelCase : str =UNetaDModel.from_pretrained(lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =DDIMScheduler() __UpperCamelCase : List[Any] =DDIMPipeline(unet=lowerCamelCase__ , scheduler=lowerCamelCase__ ) ddim.to(lowerCamelCase__ ) ddim.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Optional[int] =torch.manual_seed(0 ) __UpperCamelCase : List[str] =ddim(generator=lowerCamelCase__ , eta=0.0 , output_type='numpy' ).images __UpperCamelCase : Union[str, Any] =image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __UpperCamelCase : str =np.array([0.1_723, 0.1_617, 0.1_600, 0.1_626, 0.1_497, 0.1_513, 0.1_505, 0.1_442, 0.1_453] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[Any] ='google/ddpm-ema-bedroom-256' __UpperCamelCase : Any =UNetaDModel.from_pretrained(lowerCamelCase__ ) __UpperCamelCase : int =DDIMScheduler.from_pretrained(lowerCamelCase__ ) __UpperCamelCase : Dict =DDIMPipeline(unet=lowerCamelCase__ , scheduler=lowerCamelCase__ ) ddpm.to(lowerCamelCase__ ) ddpm.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Tuple =torch.manual_seed(0 ) __UpperCamelCase : Union[str, Any] =ddpm(generator=lowerCamelCase__ , output_type='numpy' ).images __UpperCamelCase : Tuple =image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) __UpperCamelCase : Optional[Any] =np.array([0.0_060, 0.0_201, 0.0_344, 0.0_024, 0.0_018, 0.0_002, 0.0_022, 0.0_000, 0.0_069] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
71
'''simple docstring''' # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import doctest import sys import warnings from os.path import abspath, dirname, join import _pytest from transformers.testing_utils import HfDoctestModule, HfDocTestParser # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. UpperCamelCase = abspath(join(dirname(__file__), '''src''')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='''ignore''', category=FutureWarning) def SCREAMING_SNAKE_CASE( __lowercase ) -> Optional[Any]: config.addinivalue_line( '''markers''' , '''is_pt_tf_cross_test: mark test to run only when PT and TF interactions are tested''' ) config.addinivalue_line( '''markers''' , '''is_pt_flax_cross_test: mark test to run only when PT and FLAX interactions are tested''' ) config.addinivalue_line('''markers''' , '''is_pipeline_test: mark test to run only when pipelines are tested''' ) config.addinivalue_line('''markers''' , '''is_staging_test: mark test to run only in the staging environment''' ) config.addinivalue_line('''markers''' , '''accelerate_tests: mark test that require accelerate''' ) config.addinivalue_line('''markers''' , '''tool_tests: mark the tool tests that are run on their specific schedule''' ) def SCREAMING_SNAKE_CASE( __lowercase ) -> Optional[int]: from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(__lowercase ) def SCREAMING_SNAKE_CASE( __lowercase ) -> Tuple: from transformers.testing_utils import pytest_terminal_summary_main A: Optional[int] = terminalreporter.config.getoption('''--make-reports''' ) if make_reports: pytest_terminal_summary_main(__lowercase , id=__lowercase ) def SCREAMING_SNAKE_CASE( __lowercase , __lowercase ) -> Any: # If no tests are collected, pytest exists with code 5, which makes the CI fail. if exitstatus == 5: A: Tuple = 0 # Doctest custom flag to ignore output. UpperCamelCase = doctest.register_optionflag('''IGNORE_RESULT''') UpperCamelCase = doctest.OutputChecker class lowerCAmelCase_ ( UpperCAmelCase_ ): '''simple docstring''' def _snake_case ( self : int , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : int ) -> str: '''simple docstring''' if IGNORE_RESULT & optionflags: return True return OutputChecker.check_output(self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCamelCase = CustomOutputChecker UpperCamelCase = HfDoctestModule UpperCamelCase = HfDocTestParser
319
0
"""simple docstring""" def a__ ( ): SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = 1 while len(__lowerCAmelCase ) < 1E6: constant.append(str(__lowerCAmelCase ) ) i += 1 SCREAMING_SNAKE_CASE_ = ''''''.join(__lowerCAmelCase ) return ( int(constant[0] ) * int(constant[9] ) * int(constant[9_9] ) * int(constant[9_9_9] ) * int(constant[9_9_9_9] ) * int(constant[9_9_9_9_9] ) * int(constant[9_9_9_9_9_9] ) ) if __name__ == "__main__": print(solution())
362
import unittest from transformers.testing_utils import CaptureStdout from transformers.tools.python_interpreter import evaluate def a__ ( __UpperCamelCase ): return x + 2 class lowerCamelCase (unittest.TestCase ): """simple docstring""" def __A ( self : List[Any] ) -> int: SCREAMING_SNAKE_CASE_ = "x = 3" SCREAMING_SNAKE_CASE_ = {} SCREAMING_SNAKE_CASE_ = evaluate(__magic_name__ , {} , state=__magic_name__ ) assert result == 3 self.assertDictEqual(__magic_name__ , {"x": 3} ) SCREAMING_SNAKE_CASE_ = "x = y" SCREAMING_SNAKE_CASE_ = {"y": 5} SCREAMING_SNAKE_CASE_ = evaluate(__magic_name__ , {} , state=__magic_name__ ) # evaluate returns the value of the last assignment. assert result == 5 self.assertDictEqual(__magic_name__ , {"x": 5, "y": 5} ) def __A ( self : Union[str, Any] ) -> str: SCREAMING_SNAKE_CASE_ = "y = add_two(x)" SCREAMING_SNAKE_CASE_ = {"x": 3} SCREAMING_SNAKE_CASE_ = evaluate(__magic_name__ , {"add_two": add_two} , state=__magic_name__ ) assert result == 5 self.assertDictEqual(__magic_name__ , {"x": 3, "y": 5} ) # Won't work without the tool with CaptureStdout() as out: SCREAMING_SNAKE_CASE_ = evaluate(__magic_name__ , {} , state=__magic_name__ ) assert result is None assert "tried to execute add_two" in out.out def __A ( self : List[str] ) -> int: SCREAMING_SNAKE_CASE_ = "x = 3" SCREAMING_SNAKE_CASE_ = {} SCREAMING_SNAKE_CASE_ = evaluate(__magic_name__ , {} , state=__magic_name__ ) assert result == 3 self.assertDictEqual(__magic_name__ , {"x": 3} ) def __A ( self : Optional[Any] ) -> str: SCREAMING_SNAKE_CASE_ = "test_dict = {'x': x, 'y': add_two(x)}" SCREAMING_SNAKE_CASE_ = {"x": 3} SCREAMING_SNAKE_CASE_ = evaluate(__magic_name__ , {"add_two": add_two} , state=__magic_name__ ) self.assertDictEqual(__magic_name__ , {"x": 3, "y": 5} ) self.assertDictEqual(__magic_name__ , {"x": 3, "test_dict": {"x": 3, "y": 5}} ) def __A ( self : Optional[int] ) -> List[str]: SCREAMING_SNAKE_CASE_ = "x = 3\ny = 5" SCREAMING_SNAKE_CASE_ = {} SCREAMING_SNAKE_CASE_ = evaluate(__magic_name__ , {} , state=__magic_name__ ) # evaluate returns the value of the last assignment. assert result == 5 self.assertDictEqual(__magic_name__ , {"x": 3, "y": 5} ) def __A ( self : Any ) -> List[str]: SCREAMING_SNAKE_CASE_ = "text = f'This is x: {x}.'" SCREAMING_SNAKE_CASE_ = {"x": 3} SCREAMING_SNAKE_CASE_ = evaluate(__magic_name__ , {} , state=__magic_name__ ) # evaluate returns the value of the last assignment. assert result == "This is x: 3." self.assertDictEqual(__magic_name__ , {"x": 3, "text": "This is x: 3."} ) def __A ( self : int ) -> Tuple: SCREAMING_SNAKE_CASE_ = "if x <= 3:\n y = 2\nelse:\n y = 5" SCREAMING_SNAKE_CASE_ = {"x": 3} SCREAMING_SNAKE_CASE_ = evaluate(__magic_name__ , {} , state=__magic_name__ ) # evaluate returns the value of the last assignment. assert result == 2 self.assertDictEqual(__magic_name__ , {"x": 3, "y": 2} ) SCREAMING_SNAKE_CASE_ = {"x": 8} SCREAMING_SNAKE_CASE_ = evaluate(__magic_name__ , {} , state=__magic_name__ ) # evaluate returns the value of the last assignment. assert result == 5 self.assertDictEqual(__magic_name__ , {"x": 8, "y": 5} ) def __A ( self : str ) -> str: SCREAMING_SNAKE_CASE_ = "test_list = [x, add_two(x)]" SCREAMING_SNAKE_CASE_ = {"x": 3} SCREAMING_SNAKE_CASE_ = evaluate(__magic_name__ , {"add_two": add_two} , state=__magic_name__ ) self.assertListEqual(__magic_name__ , [3, 5] ) self.assertDictEqual(__magic_name__ , {"x": 3, "test_list": [3, 5]} ) def __A ( self : Union[str, Any] ) -> List[Any]: SCREAMING_SNAKE_CASE_ = "y = x" SCREAMING_SNAKE_CASE_ = {"x": 3} SCREAMING_SNAKE_CASE_ = evaluate(__magic_name__ , {} , state=__magic_name__ ) assert result == 3 self.assertDictEqual(__magic_name__ , {"x": 3, "y": 3} ) def __A ( self : Tuple ) -> List[Any]: SCREAMING_SNAKE_CASE_ = "test_list = [x, add_two(x)]\ntest_list[1]" SCREAMING_SNAKE_CASE_ = {"x": 3} SCREAMING_SNAKE_CASE_ = evaluate(__magic_name__ , {"add_two": add_two} , state=__magic_name__ ) assert result == 5 self.assertDictEqual(__magic_name__ , {"x": 3, "test_list": [3, 5]} ) SCREAMING_SNAKE_CASE_ = "test_dict = {'x': x, 'y': add_two(x)}\ntest_dict['y']" SCREAMING_SNAKE_CASE_ = {"x": 3} SCREAMING_SNAKE_CASE_ = evaluate(__magic_name__ , {"add_two": add_two} , state=__magic_name__ ) assert result == 5 self.assertDictEqual(__magic_name__ , {"x": 3, "test_dict": {"x": 3, "y": 5}} ) def __A ( self : Tuple ) -> Any: SCREAMING_SNAKE_CASE_ = "x = 0\nfor i in range(3):\n x = i" SCREAMING_SNAKE_CASE_ = {} SCREAMING_SNAKE_CASE_ = evaluate(__magic_name__ , {"range": range} , state=__magic_name__ ) assert result == 2 self.assertDictEqual(__magic_name__ , {"x": 2, "i": 2} )
305
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available A__ = { '''configuration_gpt_neo''': ['''GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GPTNeoConfig''', '''GPTNeoOnnxConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ = [ '''GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GPTNeoForCausalLM''', '''GPTNeoForQuestionAnswering''', '''GPTNeoForSequenceClassification''', '''GPTNeoForTokenClassification''', '''GPTNeoModel''', '''GPTNeoPreTrainedModel''', '''load_tf_weights_in_gpt_neo''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ = [ '''FlaxGPTNeoForCausalLM''', '''FlaxGPTNeoModel''', '''FlaxGPTNeoPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_gpt_neo import GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoConfig, GPTNeoOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neo import ( GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoForCausalLM, GPTNeoForQuestionAnswering, GPTNeoForSequenceClassification, GPTNeoForTokenClassification, GPTNeoModel, GPTNeoPreTrainedModel, load_tf_weights_in_gpt_neo, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_gpt_neo import FlaxGPTNeoForCausalLM, FlaxGPTNeoModel, FlaxGPTNeoPreTrainedModel else: import sys A__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
230
# tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. A__ = abspath(join(dirname(dirname(dirname(__file__))), '''src''')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='''ignore''', category=FutureWarning) def _lowerCAmelCase ( __lowerCAmelCase ) -> str: """simple docstring""" from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(__lowerCAmelCase ) def _lowerCAmelCase ( __lowerCAmelCase ) -> List[Any]: """simple docstring""" from transformers.testing_utils import pytest_terminal_summary_main snake_case__ : Dict = terminalreporter.config.getoption('''--make-reports''' ) if make_reports: pytest_terminal_summary_main(__lowerCAmelCase , id=__lowerCAmelCase )
230
1
import socket def SCREAMING_SNAKE_CASE ( ) -> List[str]: UpperCamelCase__ : int = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) UpperCamelCase__ : Optional[int] = socket.gethostname() UpperCamelCase__ : List[Any] = 1_2312 sock.connect((host, port) ) sock.send(B"Hello server!" ) with open("Received_file" , "wb" ) as out_file: print("File opened" ) print("Receiving data..." ) while True: UpperCamelCase__ : int = sock.recv(1024 ) if not data: break out_file.write(__lowerCAmelCase ) print("Successfully received the file" ) sock.close() print("Connection closed" ) if __name__ == "__main__": main()
196
def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase ) -> Any: # "extended trapezoidal rule" # int(f) = dx/2 * (f1 + 2f2 + ... + fn) UpperCamelCase__ : int = (boundary[1] - boundary[0]) / steps UpperCamelCase__ : Optional[Any] = boundary[0] UpperCamelCase__ : List[Any] = boundary[1] UpperCamelCase__ : List[Any] = make_points(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) UpperCamelCase__ : int = 0.0 y += (h / 2.0) * f(__lowerCAmelCase ) for i in x_i: # print(i) y += h * f(__lowerCAmelCase ) y += (h / 2.0) * f(__lowerCAmelCase ) return y def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> List[str]: UpperCamelCase__ : Optional[int] = a + h while x < (b - h): yield x UpperCamelCase__ : Union[str, Any] = x + h def SCREAMING_SNAKE_CASE ( __lowerCAmelCase ) -> Any: # enter your function here UpperCamelCase__ : Dict = (x - 0) * (x - 0) return y def SCREAMING_SNAKE_CASE ( ) -> Dict: UpperCamelCase__ : List[Any] = 0.0 # Lower bound of integration UpperCamelCase__ : Tuple = 1.0 # Upper bound of integration UpperCamelCase__ : Any = 1_0.0 # define number of steps or resolution UpperCamelCase__ : List[str] = [a, b] # define boundary of integration UpperCamelCase__ : Any = method_a(__lowerCAmelCase , __lowerCAmelCase ) print(f'y = {y}' ) if __name__ == "__main__": main()
196
1
import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow if is_torch_available(): import torch from transformers import XLMRobertaModel @require_sentencepiece @require_tokenizers @require_torch class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def snake_case_ ( self : List[str] ): __lowercase : str = XLMRobertaModel.from_pretrained('''xlm-roberta-base''' ) __lowercase : List[Any] = torch.tensor([[0, 581, 1_0269, 83, 9_9942, 136, 6_0742, 23, 70, 8_0583, 1_8276, 2]] ) # The dog is cute and lives in the garden house __lowercase : List[str] = torch.Size((1, 12, 768) ) # batch_size, sequence_length, embedding_vector_dim __lowercase : List[str] = torch.tensor( [[-0.01_01, 0.12_18, -0.08_03, 0.08_01, 0.13_27, 0.07_76, -0.12_15, 0.23_83, 0.33_38, 0.31_06, 0.03_00, 0.02_52]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): __lowercase : Optional[int] = model(_snake_case )['''last_hidden_state'''].detach() self.assertEqual(output.shape , _snake_case ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , _snake_case , atol=1E-3 ) ) @slow def snake_case_ ( self : List[Any] ): __lowercase : Union[str, Any] = XLMRobertaModel.from_pretrained('''xlm-roberta-large''' ) __lowercase : Optional[int] = torch.tensor([[0, 581, 1_0269, 83, 9_9942, 136, 6_0742, 23, 70, 8_0583, 1_8276, 2]] ) # The dog is cute and lives in the garden house __lowercase : List[Any] = torch.Size((1, 12, 1024) ) # batch_size, sequence_length, embedding_vector_dim __lowercase : Dict = torch.tensor( [[-0.06_99, -0.03_18, 0.07_05, -0.12_41, 0.09_99, -0.05_20, 0.10_04, -0.18_38, -0.47_04, 0.14_37, 0.08_21, 0.01_26]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.large') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): __lowercase : List[Any] = model(_snake_case )['''last_hidden_state'''].detach() self.assertEqual(output.shape , _snake_case ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , _snake_case , atol=1E-3 ) )
156
from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf class __lowerCAmelCase ( lowerCAmelCase_ ): """simple docstring""" A__ : Union[str, Any] = ['''image_processor'''] A__ : Any = '''SamImageProcessor''' def __init__( self : Tuple , _snake_case : Tuple ): super().__init__(_snake_case ) __lowercase : str = self.image_processor __lowercase : Any = -10 __lowercase : Dict = self.image_processor.size['''longest_edge'''] def __call__( self : Dict , _snake_case : str=None , _snake_case : Any=None , _snake_case : List[str]=None , _snake_case : Any=None , _snake_case : Optional[Union[str, TensorType]] = None , **_snake_case : List[Any] , ): __lowercase : List[str] = self.image_processor( _snake_case , return_tensors=_snake_case , **_snake_case , ) # pop arguments that are not used in the foward but used nevertheless __lowercase : Optional[int] = encoding_image_processor['''original_sizes'''] if hasattr(_snake_case , '''numpy''' ): # Checks if Torch or TF tensor __lowercase : Optional[int] = original_sizes.numpy() __lowercase , __lowercase , __lowercase : str = self._check_and_preprocess_points( input_points=_snake_case , input_labels=_snake_case , input_boxes=_snake_case , ) __lowercase : int = self._normalize_and_convert( _snake_case , _snake_case , input_points=_snake_case , input_labels=_snake_case , input_boxes=_snake_case , return_tensors=_snake_case , ) return encoding_image_processor def snake_case_ ( self : List[str] , _snake_case : int , _snake_case : Optional[int] , _snake_case : Any=None , _snake_case : Optional[Any]=None , _snake_case : Any=None , _snake_case : str="pt" , ): if input_points is not None: if len(_snake_case ) != len(_snake_case ): __lowercase : Optional[Any] = [ self._normalize_coordinates(self.target_size , _snake_case , original_sizes[0] ) for point in input_points ] else: __lowercase : List[Any] = [ self._normalize_coordinates(self.target_size , _snake_case , _snake_case ) for point, original_size in zip(_snake_case , _snake_case ) ] # check that all arrays have the same shape if not all(point.shape == input_points[0].shape for point in input_points ): if input_labels is not None: __lowercase , __lowercase : Tuple = self._pad_points_and_labels(_snake_case , _snake_case ) __lowercase : Dict = np.array(_snake_case ) if input_labels is not None: __lowercase : Dict = np.array(_snake_case ) if input_boxes is not None: if len(_snake_case ) != len(_snake_case ): __lowercase : Union[str, Any] = [ self._normalize_coordinates(self.target_size , _snake_case , original_sizes[0] , is_bounding_box=_snake_case ) for box in input_boxes ] else: __lowercase : Tuple = [ self._normalize_coordinates(self.target_size , _snake_case , _snake_case , is_bounding_box=_snake_case ) for box, original_size in zip(_snake_case , _snake_case ) ] __lowercase : Dict = np.array(_snake_case ) if input_boxes is not None: if return_tensors == "pt": __lowercase : int = torch.from_numpy(_snake_case ) # boxes batch size of 1 by default __lowercase : List[Any] = input_boxes.unsqueeze(1 ) if len(input_boxes.shape ) != 3 else input_boxes elif return_tensors == "tf": __lowercase : Dict = tf.convert_to_tensor(_snake_case ) # boxes batch size of 1 by default __lowercase : int = tf.expand_dims(_snake_case , 1 ) if len(input_boxes.shape ) != 3 else input_boxes encoding_image_processor.update({'''input_boxes''': input_boxes} ) if input_points is not None: if return_tensors == "pt": __lowercase : Tuple = torch.from_numpy(_snake_case ) # point batch size of 1 by default __lowercase : Tuple = input_points.unsqueeze(1 ) if len(input_points.shape ) != 4 else input_points elif return_tensors == "tf": __lowercase : List[Any] = tf.convert_to_tensor(_snake_case ) # point batch size of 1 by default __lowercase : Optional[int] = tf.expand_dims(_snake_case , 1 ) if len(input_points.shape ) != 4 else input_points encoding_image_processor.update({'''input_points''': input_points} ) if input_labels is not None: if return_tensors == "pt": __lowercase : int = torch.from_numpy(_snake_case ) # point batch size of 1 by default __lowercase : Any = input_labels.unsqueeze(1 ) if len(input_labels.shape ) != 3 else input_labels elif return_tensors == "tf": __lowercase : Any = tf.convert_to_tensor(_snake_case ) # point batch size of 1 by default __lowercase : Union[str, Any] = tf.expand_dims(_snake_case , 1 ) if len(input_labels.shape ) != 3 else input_labels encoding_image_processor.update({'''input_labels''': input_labels} ) return encoding_image_processor def snake_case_ ( self : int , _snake_case : Any , _snake_case : str ): __lowercase : Union[str, Any] = max([point.shape[0] for point in input_points] ) __lowercase : List[Any] = [] for i, point in enumerate(_snake_case ): if point.shape[0] != expected_nb_points: __lowercase : Optional[Any] = np.concatenate( [point, np.zeros((expected_nb_points - point.shape[0], 2) ) + self.point_pad_value] , axis=0 ) __lowercase : Tuple = np.append(input_labels[i] , [self.point_pad_value] ) processed_input_points.append(_snake_case ) __lowercase : List[Any] = processed_input_points return input_points, input_labels def snake_case_ ( self : Dict , _snake_case : int , _snake_case : np.ndarray , _snake_case : Any , _snake_case : Any=False ): __lowercase , __lowercase : Tuple = original_size __lowercase , __lowercase : Optional[Any] = self.image_processor._get_preprocess_shape(_snake_case , longest_edge=_snake_case ) __lowercase : Optional[int] = deepcopy(_snake_case ).astype(_snake_case ) if is_bounding_box: __lowercase : str = coords.reshape(-1 , 2 , 2 ) __lowercase : Dict = coords[..., 0] * (new_w / old_w) __lowercase : int = coords[..., 1] * (new_h / old_h) if is_bounding_box: __lowercase : Optional[Any] = coords.reshape(-1 , 4 ) return coords def snake_case_ ( self : List[str] , _snake_case : List[Any]=None , _snake_case : Any=None , _snake_case : int=None , ): if input_points is not None: if hasattr(_snake_case , '''numpy''' ): # Checks for TF or Torch tensor __lowercase : Tuple = input_points.numpy().tolist() if not isinstance(_snake_case , _snake_case ) or not isinstance(input_points[0] , _snake_case ): raise ValueError('''Input points must be a list of list of floating points.''' ) __lowercase : str = [np.array(_snake_case ) for input_point in input_points] else: __lowercase : str = None if input_labels is not None: if hasattr(_snake_case , '''numpy''' ): __lowercase : Any = input_labels.numpy().tolist() if not isinstance(_snake_case , _snake_case ) or not isinstance(input_labels[0] , _snake_case ): raise ValueError('''Input labels must be a list of list integers.''' ) __lowercase : List[Any] = [np.array(_snake_case ) for label in input_labels] else: __lowercase : Tuple = None if input_boxes is not None: if hasattr(_snake_case , '''numpy''' ): __lowercase : str = input_boxes.numpy().tolist() if ( not isinstance(_snake_case , _snake_case ) or not isinstance(input_boxes[0] , _snake_case ) or not isinstance(input_boxes[0][0] , _snake_case ) ): raise ValueError('''Input boxes must be a list of list of list of floating points.''' ) __lowercase : List[Any] = [np.array(_snake_case ).astype(np.floataa ) for box in input_boxes] else: __lowercase : Dict = None return input_points, input_labels, input_boxes @property def snake_case_ ( self : List[Any] ): __lowercase : List[Any] = self.image_processor.model_input_names return list(dict.fromkeys(_snake_case ) ) def snake_case_ ( self : str , *_snake_case : Union[str, Any] , **_snake_case : Dict ): return self.image_processor.post_process_masks(*_snake_case , **_snake_case )
156
1
"""simple docstring""" import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( HubertConfig, HubertForCTC, HubertModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() __UpperCamelCase : Union[str, Any] = logging.get_logger(__name__) __UpperCamelCase : Tuple = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', } def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : Dict , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Dict , _UpperCAmelCase : List[str] , _UpperCAmelCase : List[str] ): for attribute in key.split('.' ): lowerCAmelCase = getattr(_UpperCAmelCase , _UpperCAmelCase ) if weight_type is not None: lowerCAmelCase = getattr(_UpperCAmelCase , _UpperCAmelCase ).shape else: lowerCAmelCase = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": lowerCAmelCase = value elif weight_type == "weight_g": lowerCAmelCase = value elif weight_type == "weight_v": lowerCAmelCase = value elif weight_type == "bias": lowerCAmelCase = value else: lowerCAmelCase = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : Optional[int] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Union[str, Any] ): lowerCAmelCase = [] lowerCAmelCase = fairseq_model.state_dict() lowerCAmelCase = hf_model.hubert.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): lowerCAmelCase = False if "conv_layers" in name: load_conv_layer( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , hf_model.config.feat_extract_norm == 'group' , ) lowerCAmelCase = True else: for key, mapped_key in MAPPING.items(): lowerCAmelCase = 'hubert.' + mapped_key if (is_finetuned and mapped_key != 'lm_head') else mapped_key if key in name or (key.split('w2v_model.' )[-1] == name.split('.' )[0] and not is_finetuned): lowerCAmelCase = True if "*" in mapped_key: lowerCAmelCase = name.split(_UpperCAmelCase )[0].split('.' )[-2] lowerCAmelCase = mapped_key.replace('*' , _UpperCAmelCase ) if "weight_g" in name: lowerCAmelCase = 'weight_g' elif "weight_v" in name: lowerCAmelCase = 'weight_v' elif "weight" in name: lowerCAmelCase = 'weight' elif "bias" in name: lowerCAmelCase = 'bias' else: lowerCAmelCase = None set_recursively(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) continue if not is_used: unused_weights.append(_UpperCAmelCase ) logger.warning(F'Unused weights: {unused_weights}' ) def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : Any , _UpperCAmelCase : Tuple , _UpperCAmelCase : List[str] , _UpperCAmelCase : List[str] , _UpperCAmelCase : List[Any] ): lowerCAmelCase = full_name.split('conv_layers.' )[-1] lowerCAmelCase = name.split('.' ) lowerCAmelCase = int(items[0] ) lowerCAmelCase = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) lowerCAmelCase = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) lowerCAmelCase = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) lowerCAmelCase = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) lowerCAmelCase = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(_UpperCAmelCase ) @torch.no_grad() def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : Dict , _UpperCAmelCase : int , _UpperCAmelCase : Optional[Any]=None , _UpperCAmelCase : List[str]=None , _UpperCAmelCase : Any=True ): if config_path is not None: lowerCAmelCase = HubertConfig.from_pretrained(_UpperCAmelCase ) else: lowerCAmelCase = HubertConfig() if is_finetuned: if dict_path: lowerCAmelCase = Dictionary.load(_UpperCAmelCase ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq lowerCAmelCase = target_dict.pad_index lowerCAmelCase = target_dict.bos_index lowerCAmelCase = target_dict.eos_index lowerCAmelCase = len(target_dict.symbols ) lowerCAmelCase = os.path.join(_UpperCAmelCase , 'vocab.json' ) if not os.path.isdir(_UpperCAmelCase ): logger.error('--pytorch_dump_folder_path ({}) should be a directory'.format(_UpperCAmelCase ) ) return os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase ) with open(_UpperCAmelCase , 'w' , encoding='utf-8' ) as vocab_handle: json.dump(target_dict.indices , _UpperCAmelCase ) lowerCAmelCase = WavaVecaCTCTokenizer( _UpperCAmelCase , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='|' , do_lower_case=_UpperCAmelCase , ) lowerCAmelCase = True if config.feat_extract_norm == 'layer' else False lowerCAmelCase = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6000 , padding_value=0 , do_normalize=_UpperCAmelCase , return_attention_mask=_UpperCAmelCase , ) lowerCAmelCase = WavaVecaProcessor(feature_extractor=_UpperCAmelCase , tokenizer=_UpperCAmelCase ) processor.save_pretrained(_UpperCAmelCase ) lowerCAmelCase = HubertForCTC(_UpperCAmelCase ) else: lowerCAmelCase = HubertModel(_UpperCAmelCase ) if is_finetuned: lowerCAmelCase ,lowerCAmelCase ,lowerCAmelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] )} ) else: lowerCAmelCase ,lowerCAmelCase ,lowerCAmelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) lowerCAmelCase = model[0].eval() recursively_load_weights(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) hf_wavavec.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": __UpperCamelCase : List[str] = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--not_finetuned''', action='''store_true''', help='''Whether the model to convert is a fine-tuned model or not''' ) __UpperCamelCase : List[str] = parser.parse_args() convert_hubert_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
361
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) __UpperCamelCase : Any = { '''configuration_layoutlmv2''': ['''LAYOUTLMV2_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LayoutLMv2Config'''], '''processing_layoutlmv2''': ['''LayoutLMv2Processor'''], '''tokenization_layoutlmv2''': ['''LayoutLMv2Tokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : Tuple = ['''LayoutLMv2TokenizerFast'''] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : Optional[int] = ['''LayoutLMv2FeatureExtractor'''] __UpperCamelCase : Optional[int] = ['''LayoutLMv2ImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : Any = [ '''LAYOUTLMV2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''LayoutLMv2ForQuestionAnswering''', '''LayoutLMv2ForSequenceClassification''', '''LayoutLMv2ForTokenClassification''', '''LayoutLMv2Layer''', '''LayoutLMv2Model''', '''LayoutLMv2PreTrainedModel''', ] if TYPE_CHECKING: from .configuration_layoutlmva import LAYOUTLMV2_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor, LayoutLMvaImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV2_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaLayer, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) else: import sys __UpperCamelCase : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
309
0
"""simple docstring""" import sys def lowercase ( __snake_case : Dict ): lowercase_ : int = len(__snake_case ) lowercase_ : Tuple = [[0 for x in range(__snake_case )] for x in range(__snake_case )] lowercase_ : int = [[0 for x in range(__snake_case )] for x in range(__snake_case )] for chain_length in range(2 , __snake_case ): for a in range(1 , n - chain_length + 1 ): lowercase_ : int = a + chain_length - 1 lowercase_ : str = sys.maxsize for c in range(__snake_case , __snake_case ): lowercase_ : int = ( matrix[a][c] + matrix[c + 1][b] + array[a - 1] * array[c] * array[b] ) if cost < matrix[a][b]: lowercase_ : Optional[int] = cost lowercase_ : str = c return matrix, sol def lowercase ( __snake_case : Any , __snake_case : List[str] , __snake_case : Optional[int] ): if i == j: print('''A''' + str(__snake_case ) , end=''' ''' ) else: print('''(''' , end=''' ''' ) print_optiomal_solution(__snake_case , __snake_case , optimal_solution[i][j] ) print_optiomal_solution(__snake_case , optimal_solution[i][j] + 1 , __snake_case ) print(''')''' , end=''' ''' ) def lowercase ( ): lowercase_ : Optional[int] = [3_0, 3_5, 1_5, 5, 1_0, 2_0, 2_5] lowercase_ : Optional[int] = len(__snake_case ) # Size of matrix created from above array will be # 30*35 35*15 15*5 5*10 10*20 20*25 lowercase_ , lowercase_ : Optional[Any] = matrix_chain_order(__snake_case ) print('''No. of Operation required: ''' + str(matrix[1][n - 1] ) ) print_optiomal_solution(__snake_case , 1 , n - 1 ) if __name__ == "__main__": main()
33
"""simple docstring""" from __future__ import annotations a_ = [ [-1, 0], # left [0, -1], # down [1, 0], # right [0, 1], # up ] def __UpperCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): __lowercase : Union[str, Any] = [ [0 for col in range(len(grid[0] ) )] for row in range(len(__UpperCamelCase ) ) ] # the reference grid __lowercase : Optional[int] = 1 __lowercase : str = [ [0 for col in range(len(grid[0] ) )] for row in range(len(__UpperCamelCase ) ) ] # the action grid __lowercase : List[str] = init[0] __lowercase : Optional[Any] = init[1] __lowercase : int = 0 __lowercase : List[Any] = g + heuristic[x][y] # cost from starting cell to destination cell __lowercase : Optional[Any] = [[f, g, x, y]] __lowercase : Union[str, Any] = False # flag that is set when search is complete __lowercase : List[Any] = False # flag set if we can't find expand while not found and not resign: if len(__UpperCamelCase ) == 0: raise ValueError('''Algorithm is unable to find solution''' ) else: # to choose the least costliest action so as to move closer to the goal cell.sort() cell.reverse() __lowercase : str = cell.pop() __lowercase : List[Any] = next_cell[2] __lowercase : Optional[int] = next_cell[3] __lowercase : Dict = next_cell[1] if x == goal[0] and y == goal[1]: __lowercase : List[Any] = True else: for i in range(len(__UpperCamelCase ) ): # to try out different valid actions __lowercase : Union[str, Any] = x + DIRECTIONS[i][0] __lowercase : Optional[int] = y + DIRECTIONS[i][1] if xa >= 0 and xa < len(__UpperCamelCase ) and ya >= 0 and ya < len(grid[0] ): if closed[xa][ya] == 0 and grid[xa][ya] == 0: __lowercase : str = g + cost __lowercase : Optional[int] = ga + heuristic[xa][ya] cell.append([fa, ga, xa, ya] ) __lowercase : Dict = 1 __lowercase : List[Any] = i __lowercase : Dict = [] __lowercase : List[Any] = goal[0] __lowercase : Tuple = goal[1] invpath.append([x, y] ) # we get the reverse path from here while x != init[0] or y != init[1]: __lowercase : Any = x - DIRECTIONS[action[x][y]][0] __lowercase : Dict = y - DIRECTIONS[action[x][y]][1] __lowercase : List[Any] = xa __lowercase : Optional[Any] = ya invpath.append([x, y] ) __lowercase : Optional[int] = [] for i in range(len(__UpperCamelCase ) ): path.append(invpath[len(__UpperCamelCase ) - 1 - i] ) return path, action if __name__ == "__main__": a_ = [ [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 1, 0], [0, 0, 0, 0, 1, 0], ] a_ = [0, 0] # all coordinates are given in format [y,x] a_ = [len(grid) - 1, len(grid[0]) - 1] a_ = 1 # the cost map which pushes the path closer to the goal a_ = [[0 for row in range(len(grid[0]))] for col in range(len(grid))] for i in range(len(grid)): for j in range(len(grid[0])): a_ = abs(i - goal[0]) + abs(j - goal[1]) if grid[i][j] == 1: # added extra penalty in the heuristic map a_ = 9_9 a_ , a_ = search(grid, init, goal, cost, heuristic) print('ACTION MAP') for i in range(len(action)): print(action[i]) for i in range(len(path)): print(path[i])
249
0
"""simple docstring""" # Note: if you intend to run this script make sure you look under scripts/fsmt/ # to locate the appropriate script to do the work correctly. There is a set of scripts to: # - download and prepare data and run the conversion script # - perform eval to get the best hparam into the config # - generate model_cards - useful if you have multiple models from the same paper import argparse import json import os import re from collections import OrderedDict from os.path import basename, dirname import fairseq import torch from fairseq import hub_utils from fairseq.data.dictionary import Dictionary from transformers import FSMTConfig, FSMTForConditionalGeneration from transformers.models.fsmt.tokenization_fsmt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() snake_case_ = 2 # based on the results of a search on a range of `num_beams`, `length_penalty` and `early_stopping` # values against wmt19 test data to obtain the best BLEU scores, we will use the following defaults: # # * `num_beams`: 5 (higher scores better, but requires more memory/is slower, can be adjusted by users) # * `early_stopping`: `False` consistently scored better # * `length_penalty` varied, so will assign the best one depending on the model snake_case_ = { # fairseq: """wmt19-ru-en""": {"""length_penalty""": 1.1}, """wmt19-en-ru""": {"""length_penalty""": 1.15}, """wmt19-en-de""": {"""length_penalty""": 1.0}, """wmt19-de-en""": {"""length_penalty""": 1.1}, # allenai: """wmt16-en-de-dist-12-1""": {"""length_penalty""": 0.6}, """wmt16-en-de-dist-6-1""": {"""length_penalty""": 0.6}, """wmt16-en-de-12-1""": {"""length_penalty""": 0.8}, """wmt19-de-en-6-6-base""": {"""length_penalty""": 0.6}, """wmt19-de-en-6-6-big""": {"""length_penalty""": 0.6}, } # this remaps the different models to their organization names snake_case_ = {} for m in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: snake_case_ = """facebook""" for m in [ "wmt16-en-de-dist-12-1", "wmt16-en-de-dist-6-1", "wmt16-en-de-12-1", "wmt19-de-en-6-6-base", "wmt19-de-en-6-6-big", ]: snake_case_ = """allenai""" def _lowerCAmelCase ( lowercase_ ): # (1) remove word breaking symbol, (2) add word ending symbol where the word is not broken up, # e.g.: d = {'le@@': 5, 'tt@@': 6, 'er': 7} => {'le': 5, 'tt': 6, 'er</w>': 7} UpperCAmelCase = dict((re.sub(R'@@$' , '' , lowercase_ ), v) if k.endswith('@@' ) else (re.sub(R'$' , '</w>' , lowercase_ ), v) for k, v in d.items() ) UpperCAmelCase = '<s> <pad> </s> <unk>'.split() # restore the special tokens for k in keep_keys: del da[F"""{k}</w>"""] UpperCAmelCase = d[k] # restore return da def _lowerCAmelCase ( lowercase_ , lowercase_ ): # prep assert os.path.exists(lowercase_ ) os.makedirs(lowercase_ , exist_ok=lowercase_ ) print(F"""Writing results to {pytorch_dump_folder_path}""" ) # handle various types of models UpperCAmelCase = basename(lowercase_ ) UpperCAmelCase = dirname(lowercase_ ) UpperCAmelCase = fairseq.model_parallel.models.transformer.ModelParallelTransformerModel UpperCAmelCase = cls.hub_models() UpperCAmelCase = {'bpe': 'fastbpe', 'tokenizer': 'moses'} UpperCAmelCase = '.' # note: since the model dump is old, fairseq has upgraded its model some # time later, and it does a whole lot of rewrites and splits on the saved # weights, therefore we can't use torch.load() directly on the model file. # see: upgrade_state_dict(state_dict) in fairseq_model.py print(F"""using checkpoint {checkpoint_file}""" ) UpperCAmelCase = hub_utils.from_pretrained( lowercase_ , lowercase_ , lowercase_ , archive_map=lowercase_ , **lowercase_ ) UpperCAmelCase = vars(chkpt['args']['model'] ) UpperCAmelCase = args['source_lang'] UpperCAmelCase = args['target_lang'] UpperCAmelCase = dirname(lowercase_ ) UpperCAmelCase = basename(lowercase_ ) # dicts UpperCAmelCase = os.path.join(lowercase_ , F"""dict.{src_lang}.txt""" ) UpperCAmelCase = os.path.join(lowercase_ , F"""dict.{tgt_lang}.txt""" ) UpperCAmelCase = Dictionary.load(lowercase_ ) UpperCAmelCase = rewrite_dict_keys(src_dict.indices ) UpperCAmelCase = len(lowercase_ ) UpperCAmelCase = os.path.join(lowercase_ , 'vocab-src.json' ) print(F"""Generating {src_vocab_file} of {src_vocab_size} of {src_lang} records""" ) with open(lowercase_ , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(lowercase_ , ensure_ascii=lowercase_ , indent=lowercase_ ) ) # detect whether this is a do_lower_case situation, which can be derived by checking whether we # have at least one uppercase letter in the source vocab UpperCAmelCase = True for k in src_vocab.keys(): if not k.islower(): UpperCAmelCase = False break UpperCAmelCase = Dictionary.load(lowercase_ ) UpperCAmelCase = rewrite_dict_keys(tgt_dict.indices ) UpperCAmelCase = len(lowercase_ ) UpperCAmelCase = os.path.join(lowercase_ , 'vocab-tgt.json' ) print(F"""Generating {tgt_vocab_file} of {tgt_vocab_size} of {tgt_lang} records""" ) with open(lowercase_ , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(lowercase_ , ensure_ascii=lowercase_ , indent=lowercase_ ) ) # merges_file (bpecodes) UpperCAmelCase = os.path.join(lowercase_ , VOCAB_FILES_NAMES['merges_file'] ) for fn in ["bpecodes", "code"]: # older fairseq called the merges file "code" UpperCAmelCase = os.path.join(lowercase_ , lowercase_ ) if os.path.exists(lowercase_ ): break with open(lowercase_ , encoding='utf-8' ) as fin: UpperCAmelCase = fin.read() UpperCAmelCase = re.sub(R' \d+$' , '' , lowercase_ , 0 , re.M ) # remove frequency number print(F"""Generating {merges_file}""" ) with open(lowercase_ , 'w' , encoding='utf-8' ) as fout: fout.write(lowercase_ ) # model config UpperCAmelCase = os.path.join(lowercase_ , 'config.json' ) # validate bpe/tokenizer config, as currently it's hardcoded to moses+fastbpe - # may have to modify the tokenizer if a different type is used by a future model assert args["bpe"] == "fastbpe", F"""need to extend tokenizer to support bpe={args['bpe']}""" assert args["tokenizer"] == "moses", F"""need to extend tokenizer to support bpe={args['tokenizer']}""" UpperCAmelCase = { 'architectures': ['FSMTForConditionalGeneration'], 'model_type': 'fsmt', 'activation_dropout': args['activation_dropout'], 'activation_function': 'relu', 'attention_dropout': args['attention_dropout'], 'd_model': args['decoder_embed_dim'], 'dropout': args['dropout'], 'init_std': 0.0_2, 'max_position_embeddings': args['max_source_positions'], 'num_hidden_layers': args['encoder_layers'], 'src_vocab_size': src_vocab_size, 'tgt_vocab_size': tgt_vocab_size, 'langs': [src_lang, tgt_lang], 'encoder_attention_heads': args['encoder_attention_heads'], 'encoder_ffn_dim': args['encoder_ffn_embed_dim'], 'encoder_layerdrop': args['encoder_layerdrop'], 'encoder_layers': args['encoder_layers'], 'decoder_attention_heads': args['decoder_attention_heads'], 'decoder_ffn_dim': args['decoder_ffn_embed_dim'], 'decoder_layerdrop': args['decoder_layerdrop'], 'decoder_layers': args['decoder_layers'], 'bos_token_id': 0, 'pad_token_id': 1, 'eos_token_id': 2, 'is_encoder_decoder': True, 'scale_embedding': not args['no_scale_embedding'], 'tie_word_embeddings': args['share_all_embeddings'], } # good hparam defaults to start with UpperCAmelCase = 5 UpperCAmelCase = False if model_dir in best_score_hparams and "length_penalty" in best_score_hparams[model_dir]: UpperCAmelCase = best_score_hparams[model_dir]['length_penalty'] else: UpperCAmelCase = 1.0 print(F"""Generating {fsmt_model_config_file}""" ) with open(lowercase_ , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(lowercase_ , ensure_ascii=lowercase_ , indent=lowercase_ ) ) # tokenizer config UpperCAmelCase = os.path.join(lowercase_ , lowercase_ ) UpperCAmelCase = { 'langs': [src_lang, tgt_lang], 'model_max_length': 1024, 'do_lower_case': do_lower_case, } print(F"""Generating {fsmt_tokenizer_config_file}""" ) with open(lowercase_ , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(lowercase_ , ensure_ascii=lowercase_ , indent=lowercase_ ) ) # model UpperCAmelCase = chkpt['models'][0] UpperCAmelCase = model.state_dict() # rename keys to start with 'model.' UpperCAmelCase = OrderedDict(('model.' + k, v) for k, v in model_state_dict.items() ) # remove unneeded keys UpperCAmelCase = [ 'model.model', 'model.encoder.version', 'model.decoder.version', 'model.encoder_embed_tokens.weight', 'model.decoder_embed_tokens.weight', 'model.encoder.embed_positions._float_tensor', 'model.decoder.embed_positions._float_tensor', ] for k in ignore_keys: model_state_dict.pop(lowercase_ , lowercase_ ) UpperCAmelCase = FSMTConfig.from_pretrained(lowercase_ ) UpperCAmelCase = FSMTForConditionalGeneration(lowercase_ ) # check that it loads ok model_new.load_state_dict(lowercase_ , strict=lowercase_ ) # save UpperCAmelCase = os.path.join(lowercase_ , lowercase_ ) print(F"""Generating {pytorch_weights_dump_path}""" ) torch.save(lowercase_ , lowercase_ ) print('Conversion is done!' ) print('\nLast step is to upload the files to s3' ) print(F"""cd {data_root}""" ) print(F"""transformers-cli upload {model_dir}""" ) if __name__ == "__main__": snake_case_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--fsmt_checkpoint_path""", default=None, type=str, required=True, help=( """Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts,""" """ bpecodes, etc.""" ), ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) snake_case_ = parser.parse_args() convert_fsmt_checkpoint_to_pytorch(args.fsmt_checkpoint_path, args.pytorch_dump_folder_path)
181
"""simple docstring""" def _lowerCAmelCase ( lowercase_ ): if not isinstance(lowercase_ , lowercase_ ): UpperCAmelCase = F"""Input value of [number={number}] must be an integer""" raise TypeError(lowercase_ ) if number < 1: UpperCAmelCase = F"""Input value of [number={number}] must be > 0""" raise ValueError(lowercase_ ) UpperCAmelCase = 1 for i in range(1 , lowercase_ ): current_number *= 4 * i - 2 current_number //= i + 1 return current_number if __name__ == "__main__": import doctest doctest.testmod()
181
1
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import is_speech_available from transformers.testing_utils import require_torch, require_torchaudio from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import SpeechaTextFeatureExtractor lowerCamelCase__ = random.Random() def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase=1.0 , _UpperCamelCase=None , _UpperCamelCase=None ): if rng is None: __lowerCAmelCase : int = global_rng __lowerCAmelCase : int = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class A__ ( unittest.TestCase): def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=7 , _SCREAMING_SNAKE_CASE=4_00 , _SCREAMING_SNAKE_CASE=20_00 , _SCREAMING_SNAKE_CASE=24 , _SCREAMING_SNAKE_CASE=24 , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=1_60_00 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , ): __lowerCAmelCase : List[Any] = parent __lowerCAmelCase : str = batch_size __lowerCAmelCase : Union[str, Any] = min_seq_length __lowerCAmelCase : Union[str, Any] = max_seq_length __lowerCAmelCase : int = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __lowerCAmelCase : List[Any] = feature_size __lowerCAmelCase : str = num_mel_bins __lowerCAmelCase : Union[str, Any] = padding_value __lowerCAmelCase : Optional[int] = sampling_rate __lowerCAmelCase : Dict = return_attention_mask __lowerCAmelCase : Tuple = do_normalize def __lowerCamelCase ( self ): return { "feature_size": self.feature_size, "num_mel_bins": self.num_mel_bins, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=False ): def _flatten(_SCREAMING_SNAKE_CASE ): return list(itertools.chain(*_SCREAMING_SNAKE_CASE ) ) if equal_length: __lowerCAmelCase : List[str] = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size __lowerCAmelCase : Union[str, Any] = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __lowerCAmelCase : str = [np.asarray(_SCREAMING_SNAKE_CASE ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class A__ ( _lowerCamelCase , unittest.TestCase): A_ : List[Any] = SpeechaTextFeatureExtractor if is_speech_available() else None def __lowerCamelCase ( self ): __lowerCAmelCase : Optional[Any] = SpeechaTextFeatureExtractionTester(self ) def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE ): self.assertTrue(np.all(np.mean(_SCREAMING_SNAKE_CASE , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(_SCREAMING_SNAKE_CASE , axis=0 ) - 1 ) < 1E-3 ) ) def __lowerCamelCase ( self ): # Tests that all call wrap to encode_plus and batch_encode_plus __lowerCAmelCase : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __lowerCAmelCase : Any = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] __lowerCAmelCase : Union[str, Any] = [np.asarray(_SCREAMING_SNAKE_CASE ) for speech_input in speech_inputs] # Test feature size __lowerCAmelCase : str = feature_extractor(_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE , return_tensors='np' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.feature_size ) # Test not batched input __lowerCAmelCase : List[Any] = feature_extractor(speech_inputs[0] , return_tensors='np' ).input_features __lowerCAmelCase : Optional[int] = feature_extractor(np_speech_inputs[0] , return_tensors='np' ).input_features self.assertTrue(np.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=1E-3 ) ) # Test batched __lowerCAmelCase : List[Any] = feature_extractor(_SCREAMING_SNAKE_CASE , return_tensors='np' ).input_features __lowerCAmelCase : Tuple = feature_extractor(_SCREAMING_SNAKE_CASE , return_tensors='np' ).input_features for enc_seq_a, enc_seq_a in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): self.assertTrue(np.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. __lowerCAmelCase : int = [floats_list((1, x) )[0] for x in (8_00, 8_00, 8_00)] __lowerCAmelCase : Optional[Any] = np.asarray(_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Optional[int] = feature_extractor(_SCREAMING_SNAKE_CASE , return_tensors='np' ).input_features __lowerCAmelCase : str = feature_extractor(_SCREAMING_SNAKE_CASE , return_tensors='np' ).input_features for enc_seq_a, enc_seq_a in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): self.assertTrue(np.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=1E-3 ) ) def __lowerCamelCase ( self ): __lowerCAmelCase : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Optional[int] = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] __lowerCAmelCase : List[Any] = ['longest', 'max_length', 'do_not_pad'] __lowerCAmelCase : Union[str, Any] = [None, 16, None] for max_length, padding in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): __lowerCAmelCase : Optional[Any] = feature_extractor( _SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE , max_length=_SCREAMING_SNAKE_CASE , return_attention_mask=_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Union[str, Any] = inputs.input_features __lowerCAmelCase : List[Any] = inputs.attention_mask __lowerCAmelCase : Union[str, Any] = [np.sum(_SCREAMING_SNAKE_CASE ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def __lowerCamelCase ( self ): __lowerCAmelCase : Optional[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Any = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] __lowerCAmelCase : List[str] = ['longest', 'max_length', 'do_not_pad'] __lowerCAmelCase : Any = [None, 16, None] for max_length, padding in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): __lowerCAmelCase : str = feature_extractor( _SCREAMING_SNAKE_CASE , max_length=_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE , return_tensors='np' , return_attention_mask=_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Dict = inputs.input_features __lowerCAmelCase : str = inputs.attention_mask __lowerCAmelCase : Union[str, Any] = [np.sum(_SCREAMING_SNAKE_CASE ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self.assertTrue(input_features[0][fbank_feat_lengths[0] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self.assertTrue(input_features[0][fbank_feat_lengths[1] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def __lowerCamelCase ( self ): __lowerCAmelCase : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Union[str, Any] = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] __lowerCAmelCase : Optional[int] = feature_extractor( _SCREAMING_SNAKE_CASE , padding='max_length' , max_length=4 , truncation=_SCREAMING_SNAKE_CASE , return_tensors='np' , return_attention_mask=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : int = inputs.input_features __lowerCAmelCase : Optional[Any] = inputs.attention_mask __lowerCAmelCase : Union[str, Any] = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1] ) self._check_zero_mean_unit_variance(input_features[2] ) def __lowerCamelCase ( self ): __lowerCAmelCase : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Any = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] __lowerCAmelCase : int = feature_extractor( _SCREAMING_SNAKE_CASE , padding='longest' , max_length=4 , truncation=_SCREAMING_SNAKE_CASE , return_tensors='np' , return_attention_mask=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Optional[Any] = inputs.input_features __lowerCAmelCase : List[Any] = inputs.attention_mask __lowerCAmelCase : Optional[Any] = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 4, 24) ) __lowerCAmelCase : Any = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] __lowerCAmelCase : Optional[int] = feature_extractor( _SCREAMING_SNAKE_CASE , padding='longest' , max_length=16 , truncation=_SCREAMING_SNAKE_CASE , return_tensors='np' , return_attention_mask=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : List[Any] = inputs.input_features __lowerCAmelCase : Any = inputs.attention_mask __lowerCAmelCase : Any = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 6, 24) ) def __lowerCamelCase ( self ): import torch __lowerCAmelCase : int = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : Tuple = np.random.rand(1_00 , 32 ).astype(np.floataa ) __lowerCAmelCase : Optional[int] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __lowerCAmelCase : Optional[Any] = feature_extractor.pad([{'input_features': inputs}] , return_tensors='np' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) __lowerCAmelCase : Optional[int] = feature_extractor.pad([{'input_features': inputs}] , return_tensors='pt' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE ): from datasets import load_dataset __lowerCAmelCase : str = load_dataset('hf-internal-testing/librispeech_asr_dummy' , 'clean' , split='validation' ) # automatic decoding with librispeech __lowerCAmelCase : Union[str, Any] = ds.sort('id' ).select(range(_SCREAMING_SNAKE_CASE ) )[:num_samples]['audio'] return [x["array"] for x in speech_samples] def __lowerCamelCase ( self ): # fmt: off __lowerCAmelCase : Tuple = np.array([ -1.5745, -1.7713, -1.7020, -1.6069, -1.2250, -1.1105, -0.9072, -0.8241, -1.2310, -0.8098, -0.3320, -0.4101, -0.7985, -0.4996, -0.8213, -0.9128, -1.0420, -1.1286, -1.0440, -0.7999, -0.8405, -1.2275, -1.5443, -1.4625, ] ) # fmt: on __lowerCAmelCase : List[str] = self._load_datasamples(1 ) __lowerCAmelCase : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase : str = feature_extractor(_SCREAMING_SNAKE_CASE , return_tensors='pt' ).input_features self.assertEquals(input_features.shape , (1, 5_84, 24) ) self.assertTrue(np.allclose(input_features[0, 0, :30] , _SCREAMING_SNAKE_CASE , atol=1E-4 ) )
86
"""simple docstring""" import math import sys def __lowerCAmelCase (_UpperCamelCase ): if number != int(_UpperCamelCase ): raise ValueError('the value of input must be a natural number' ) if number < 0: raise ValueError('the value of input must not be a negative number' ) if number == 0: return 1 __lowerCAmelCase : Any = [-1] * (number + 1) __lowerCAmelCase : List[Any] = 0 for i in range(1 , number + 1 ): __lowerCAmelCase : List[Any] = sys.maxsize __lowerCAmelCase : Optional[int] = int(math.sqrt(_UpperCamelCase ) ) for j in range(1 , root + 1 ): __lowerCAmelCase : Optional[Any] = 1 + answers[i - (j**2)] __lowerCAmelCase : Any = min(_UpperCamelCase , _UpperCamelCase ) __lowerCAmelCase : List[str] = answer return answers[number] if __name__ == "__main__": import doctest doctest.testmod()
86
1
import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / '''utils''')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 lowerCamelCase__ = get_tests_dir('''fixtures''') lowerCamelCase__ = get_tests_dir('''fixtures/dummy_feature_extractor_config.json''') lowerCamelCase__ = get_tests_dir('''fixtures/dummy-config.json''') class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : Tuple) -> Tuple: """simple docstring""" _UpperCamelCase = 0 def __UpperCAmelCase ( self : List[str]) -> Tuple: """simple docstring""" _UpperCamelCase = AutoFeatureExtractor.from_pretrained("facebook/wav2vec2-base-960h") self.assertIsInstance(lowercase_ , lowercase_) def __UpperCAmelCase ( self : List[Any]) -> str: """simple docstring""" _UpperCamelCase = AutoFeatureExtractor.from_pretrained(lowercase_) self.assertIsInstance(lowercase_ , lowercase_) def __UpperCAmelCase ( self : int) -> Optional[int]: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: _UpperCamelCase = WavaVecaConfig() # remove feature_extractor_type to make sure config.json alone is enough to load feature processor locally _UpperCamelCase = AutoFeatureExtractor.from_pretrained(lowercase_).to_dict() config_dict.pop("feature_extractor_type") _UpperCamelCase = WavaVecaFeatureExtractor(**lowercase_) # save in new folder model_config.save_pretrained(lowercase_) config.save_pretrained(lowercase_) _UpperCamelCase = AutoFeatureExtractor.from_pretrained(lowercase_) # make sure private variable is not incorrectly saved _UpperCamelCase = json.loads(config.to_json_string()) self.assertTrue("_processor_class" not in dict_as_saved) self.assertIsInstance(lowercase_ , lowercase_) def __UpperCAmelCase ( self : int) -> Tuple: """simple docstring""" _UpperCamelCase = AutoFeatureExtractor.from_pretrained(lowercase_) self.assertIsInstance(lowercase_ , lowercase_) def __UpperCAmelCase ( self : Any) -> Tuple: """simple docstring""" with self.assertRaisesRegex( lowercase_ , "bert-base is not a local folder and is not a valid model identifier"): _UpperCamelCase = AutoFeatureExtractor.from_pretrained("bert-base") def __UpperCAmelCase ( self : List[Any]) -> List[str]: """simple docstring""" with self.assertRaisesRegex( lowercase_ , R"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)"): _UpperCamelCase = AutoFeatureExtractor.from_pretrained(lowercase_ , revision="aaaaaa") def __UpperCAmelCase ( self : Union[str, Any]) -> Union[str, Any]: """simple docstring""" with self.assertRaisesRegex( lowercase_ , "hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json." , ): _UpperCamelCase = AutoFeatureExtractor.from_pretrained("hf-internal-testing/config-no-model") def __UpperCAmelCase ( self : Dict) -> Union[str, Any]: """simple docstring""" with self.assertRaises(lowercase_): _UpperCamelCase = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor") # If remote code is disabled, we can't load this config. with self.assertRaises(lowercase_): _UpperCamelCase = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=lowercase_) _UpperCamelCase = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=lowercase_) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor") # Test feature extractor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(lowercase_) _UpperCamelCase = AutoFeatureExtractor.from_pretrained(lowercase_ , trust_remote_code=lowercase_) self.assertEqual(reloaded_feature_extractor.__class__.__name__ , "NewFeatureExtractor") def __UpperCAmelCase ( self : Dict) -> List[Any]: """simple docstring""" try: AutoConfig.register("custom" , lowercase_) AutoFeatureExtractor.register(lowercase_ , lowercase_) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(lowercase_): AutoFeatureExtractor.register(lowercase_ , lowercase_) # Now that the config is registered, it can be used as any other config with the auto-API _UpperCamelCase = CustomFeatureExtractor.from_pretrained(lowercase_) with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(lowercase_) _UpperCamelCase = AutoFeatureExtractor.from_pretrained(lowercase_) self.assertIsInstance(lowercase_ , lowercase_) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] def __UpperCAmelCase ( self : Union[str, Any]) -> str: """simple docstring""" class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = True try: AutoConfig.register("custom" , lowercase_) AutoFeatureExtractor.register(lowercase_ , lowercase_) # If remote code is not set, the default is to use local _UpperCamelCase = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor") self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor") self.assertTrue(feature_extractor.is_local) # If remote code is disabled, we load the local one. _UpperCamelCase = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=lowercase_) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor") self.assertTrue(feature_extractor.is_local) # If remote is enabled, we load from the Hub _UpperCamelCase = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=lowercase_) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor") self.assertTrue(not hasattr(lowercase_ , "is_local")) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig]
63
import os def lowerCAmelCase__ ( ) ->Any: '''simple docstring''' with open(os.path.dirname(a__ ) + "/grid.txt" ) as f: _UpperCamelCase = [] # noqa: E741 for _ in range(20 ): l.append([int(a__ ) for x in f.readline().split()] ) _UpperCamelCase = 0 # right for i in range(20 ): for j in range(17 ): _UpperCamelCase = l[i][j] * l[i][j + 1] * l[i][j + 2] * l[i][j + 3] if temp > maximum: _UpperCamelCase = temp # down for i in range(17 ): for j in range(20 ): _UpperCamelCase = l[i][j] * l[i + 1][j] * l[i + 2][j] * l[i + 3][j] if temp > maximum: _UpperCamelCase = temp # diagonal 1 for i in range(17 ): for j in range(17 ): _UpperCamelCase = l[i][j] * l[i + 1][j + 1] * l[i + 2][j + 2] * l[i + 3][j + 3] if temp > maximum: _UpperCamelCase = temp # diagonal 2 for i in range(17 ): for j in range(3 , 20 ): _UpperCamelCase = l[i][j] * l[i + 1][j - 1] * l[i + 2][j - 2] * l[i + 3][j - 3] if temp > maximum: _UpperCamelCase = temp return maximum if __name__ == "__main__": print(solution())
63
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase : Tuple = logging.get_logger(__name__) _lowerCamelCase : Union[str, Any] = { "google/canine-s": "https://huggingface.co/google/canine-s/resolve/main/config.json", # See all CANINE models at https://huggingface.co/models?filter=canine } class __UpperCAmelCase ( lowerCamelCase__ ): UpperCamelCase = """canine""" def __init__( self : Dict, __A : str=7_6_8, __A : Tuple=1_2, __A : str=1_2, __A : List[str]=3_0_7_2, __A : Union[str, Any]="gelu", __A : Any=0.1, __A : Union[str, Any]=0.1, __A : Dict=1_6_3_8_4, __A : List[str]=1_6, __A : str=0.0_2, __A : List[str]=1E-12, __A : Optional[int]=0, __A : str=0XE000, __A : int=0XE001, __A : Any=4, __A : Tuple=4, __A : str=8, __A : Optional[int]=1_6_3_8_4, __A : Optional[Any]=1_2_8, **__A : List[Any], ): super().__init__(pad_token_id=__A, bos_token_id=__A, eos_token_id=__A, **__A ) UpperCAmelCase : List[str] = max_position_embeddings UpperCAmelCase : Any = hidden_size UpperCAmelCase : Dict = num_hidden_layers UpperCAmelCase : int = num_attention_heads UpperCAmelCase : str = intermediate_size UpperCAmelCase : Dict = hidden_act UpperCAmelCase : Optional[Any] = hidden_dropout_prob UpperCAmelCase : Optional[Any] = attention_probs_dropout_prob UpperCAmelCase : Optional[Any] = initializer_range UpperCAmelCase : Optional[int] = type_vocab_size UpperCAmelCase : Union[str, Any] = layer_norm_eps # Character config: UpperCAmelCase : Dict = downsampling_rate UpperCAmelCase : Tuple = upsampling_kernel_size UpperCAmelCase : str = num_hash_functions UpperCAmelCase : Tuple = num_hash_buckets UpperCAmelCase : Dict = local_transformer_stride
336
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) _lowerCamelCase : Tuple = { "configuration_encodec": [ "ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP", "EncodecConfig", ], "feature_extraction_encodec": ["EncodecFeatureExtractor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Optional[Any] = [ "ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST", "EncodecModel", "EncodecPreTrainedModel", ] if TYPE_CHECKING: from .configuration_encodec import ( ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP, EncodecConfig, ) from .feature_extraction_encodec import EncodecFeatureExtractor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encodec import ( ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST, EncodecModel, EncodecPreTrainedModel, ) else: import sys _lowerCamelCase : Optional[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
336
1
import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import TOKEN, USER, is_staging_test sys.path.append(str(Path(__file__).parent.parent / 'utils')) from test_module.custom_configuration import CustomConfig # noqa E402 lowerCAmelCase__ : Optional[Any] ={ 'return_dict': False, 'output_hidden_states': True, 'output_attentions': True, 'torchscript': True, 'torch_dtype': 'float16', 'use_bfloat16': True, 'tf_legacy_loss': True, 'pruned_heads': {'a': 1}, 'tie_word_embeddings': False, 'is_decoder': True, 'cross_attention_hidden_size': 1_28, 'add_cross_attention': True, 'tie_encoder_decoder': True, 'max_length': 50, 'min_length': 3, 'do_sample': True, 'early_stopping': True, 'num_beams': 3, 'num_beam_groups': 3, 'diversity_penalty': 0.5, 'temperature': 2.0, 'top_k': 10, 'top_p': 0.7, 'typical_p': 0.2, 'repetition_penalty': 0.8, 'length_penalty': 0.8, 'no_repeat_ngram_size': 5, 'encoder_no_repeat_ngram_size': 5, 'bad_words_ids': [1, 2, 3], 'num_return_sequences': 3, 'chunk_size_feed_forward': 5, 'output_scores': True, 'return_dict_in_generate': True, 'forced_bos_token_id': 2, 'forced_eos_token_id': 3, 'remove_invalid_values': True, 'architectures': ['BertModel'], 'finetuning_task': 'translation', 'id2label': {0: 'label'}, 'label2id': {'label': '0'}, 'tokenizer_class': 'BertTokenizerFast', 'prefix': 'prefix', 'bos_token_id': 6, 'pad_token_id': 7, 'eos_token_id': 8, 'sep_token_id': 9, 'decoder_start_token_id': 10, 'exponential_decay_length_penalty': (5, 1.0_1), 'suppress_tokens': [0, 1], 'begin_suppress_tokens': 2, 'task_specific_params': {'translation': 'some_params'}, 'problem_type': 'regression', } @is_staging_test class __lowercase (unittest.TestCase ): """simple docstring""" @classmethod def UpperCamelCase__ ( cls ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = TOKEN HfFolder.save_token(lowerCAmelCase__ ) @classmethod def UpperCamelCase__ ( cls ): """simple docstring""" try: delete_repo(token=cls._token , repo_id='test-config' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='valid_org/test-config-org' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='test-dynamic-config' ) except HTTPError: pass def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = BertConfig( vocab_size=9_9 , hidden_size=3_2 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=3_7 ) config.push_to_hub('test-config' , use_auth_token=self._token ) SCREAMING_SNAKE_CASE_ : Optional[int] = BertConfig.from_pretrained(F'''{USER}/test-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCAmelCase__ , getattr(lowerCAmelCase__ , lowerCAmelCase__ ) ) # Reset repo delete_repo(token=self._token , repo_id='test-config' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowerCAmelCase__ , repo_id='test-config' , push_to_hub=lowerCAmelCase__ , use_auth_token=self._token ) SCREAMING_SNAKE_CASE_ : Optional[int] = BertConfig.from_pretrained(F'''{USER}/test-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCAmelCase__ , getattr(lowerCAmelCase__ , lowerCAmelCase__ ) ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = BertConfig( vocab_size=9_9 , hidden_size=3_2 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=3_7 ) config.push_to_hub('valid_org/test-config-org' , use_auth_token=self._token ) SCREAMING_SNAKE_CASE_ : Dict = BertConfig.from_pretrained('valid_org/test-config-org' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCAmelCase__ , getattr(lowerCAmelCase__ , lowerCAmelCase__ ) ) # Reset repo delete_repo(token=self._token , repo_id='valid_org/test-config-org' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( lowerCAmelCase__ , repo_id='valid_org/test-config-org' , push_to_hub=lowerCAmelCase__ , use_auth_token=self._token ) SCREAMING_SNAKE_CASE_ : Optional[int] = BertConfig.from_pretrained('valid_org/test-config-org' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCAmelCase__ , getattr(lowerCAmelCase__ , lowerCAmelCase__ ) ) def UpperCamelCase__ ( self ): """simple docstring""" CustomConfig.register_for_auto_class() SCREAMING_SNAKE_CASE_ : Tuple = CustomConfig(attribute=4_2 ) config.push_to_hub('test-dynamic-config' , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual(config.auto_map , {'AutoConfig': 'custom_configuration.CustomConfig'} ) SCREAMING_SNAKE_CASE_ : Optional[int] = AutoConfig.from_pretrained(F'''{USER}/test-dynamic-config''' , trust_remote_code=lowerCAmelCase__ ) # Can't make an isinstance check because the new_config is from the FakeConfig class of a dynamic module self.assertEqual(new_config.__class__.__name__ , 'CustomConfig' ) self.assertEqual(new_config.attribute , 4_2 ) class __lowercase (unittest.TestCase ): """simple docstring""" def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = GPTaConfig() # attempt to modify each of int/float/bool/str config records and verify they were updated SCREAMING_SNAKE_CASE_ : Optional[int] = c.n_embd + 1 # int SCREAMING_SNAKE_CASE_ : int = c.resid_pdrop + 1.0 # float SCREAMING_SNAKE_CASE_ : Union[str, Any] = not c.scale_attn_weights # bool SCREAMING_SNAKE_CASE_ : str = c.summary_type + 'foo' # str c.update_from_string( F'''n_embd={n_embd},resid_pdrop={resid_pdrop},scale_attn_weights={scale_attn_weights},summary_type={summary_type}''' ) self.assertEqual(lowerCAmelCase__ , c.n_embd , 'mismatch for key: n_embd' ) self.assertEqual(lowerCAmelCase__ , c.resid_pdrop , 'mismatch for key: resid_pdrop' ) self.assertEqual(lowerCAmelCase__ , c.scale_attn_weights , 'mismatch for key: scale_attn_weights' ) self.assertEqual(lowerCAmelCase__ , c.summary_type , 'mismatch for key: summary_type' ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str = PretrainedConfig() SCREAMING_SNAKE_CASE_ : str = [key for key in base_config.__dict__ if key not in config_common_kwargs] # If this part of the test fails, you have arguments to addin config_common_kwargs above. self.assertListEqual( lowerCAmelCase__ , ['is_encoder_decoder', '_name_or_path', '_commit_hash', 'transformers_version'] ) SCREAMING_SNAKE_CASE_ : Optional[int] = [key for key, value in config_common_kwargs.items() if value == getattr(lowerCAmelCase__ , lowerCAmelCase__ )] if len(lowerCAmelCase__ ) > 0: raise ValueError( 'The following keys are set with the default values in' ' `test_configuration_common.config_common_kwargs` pick another value for them:' F''' {', '.join(lowerCAmelCase__ )}.''' ) def UpperCamelCase__ ( self ): """simple docstring""" with self.assertRaises(lowerCAmelCase__ ): # config is in subfolder, the following should not work without specifying the subfolder SCREAMING_SNAKE_CASE_ : List[Any] = BertConfig.from_pretrained('hf-internal-testing/tiny-random-bert-subfolder' ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = BertConfig.from_pretrained('hf-internal-testing/tiny-random-bert-subfolder' , subfolder='bert' ) self.assertIsNotNone(lowerCAmelCase__ ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = mock.Mock() SCREAMING_SNAKE_CASE_ : str = 5_0_0 SCREAMING_SNAKE_CASE_ : List[Any] = {} SCREAMING_SNAKE_CASE_ : List[str] = HTTPError SCREAMING_SNAKE_CASE_ : str = {} # Download this model to make sure it's in the cache. SCREAMING_SNAKE_CASE_ : Union[str, Any] = BertConfig.from_pretrained('hf-internal-testing/tiny-random-bert' ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch('requests.Session.request' , return_value=lowerCAmelCase__ ) as mock_head: SCREAMING_SNAKE_CASE_ : List[str] = BertConfig.from_pretrained('hf-internal-testing/tiny-random-bert' ) # This check we did call the fake head request mock_head.assert_called() def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = BertConfig.from_pretrained( 'https://huggingface.co/hf-internal-testing/tiny-random-bert/resolve/main/config.json' ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = AutoConfig.from_pretrained('bert-base-cased' ) SCREAMING_SNAKE_CASE_ : List[str] = ['config.4.0.0.json'] with tempfile.TemporaryDirectory() as tmp_dir: configuration.save_pretrained(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : str = 2 json.dump(configuration.to_dict() , open(os.path.join(lowerCAmelCase__ , 'config.4.0.0.json' ) , 'w' ) ) # This should pick the new configuration file as the version of Transformers is > 4.0.0 SCREAMING_SNAKE_CASE_ : Dict = AutoConfig.from_pretrained(lowerCAmelCase__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # Will need to be adjusted if we reach v42 and this test is still here. # Should pick the old configuration file as the version of Transformers is < 4.42.0 SCREAMING_SNAKE_CASE_ : Optional[int] = ['config.42.0.0.json'] SCREAMING_SNAKE_CASE_ : Any = 7_6_8 configuration.save_pretrained(lowerCAmelCase__ ) shutil.move(os.path.join(lowerCAmelCase__ , 'config.4.0.0.json' ) , os.path.join(lowerCAmelCase__ , 'config.42.0.0.json' ) ) SCREAMING_SNAKE_CASE_ : Any = AutoConfig.from_pretrained(lowerCAmelCase__ ) self.assertEqual(new_configuration.hidden_size , 7_6_8 ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = 'hf-internal-testing/test-two-configs' import transformers as new_transformers SCREAMING_SNAKE_CASE_ : str = 'v4.0.0' SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Union[str, Any] = new_transformers.models.auto.AutoConfig.from_pretrained( lowerCAmelCase__ , return_unused_kwargs=lowerCAmelCase__ ) self.assertEqual(new_configuration.hidden_size , 2 ) # This checks `_configuration_file` ia not kept in the kwargs by mistake. self.assertDictEqual(lowerCAmelCase__ , {} ) # Testing an older version by monkey-patching the version in the module it's used. import transformers as old_transformers SCREAMING_SNAKE_CASE_ : Optional[Any] = 'v3.0.0' SCREAMING_SNAKE_CASE_ : List[str] = old_transformers.models.auto.AutoConfig.from_pretrained(lowerCAmelCase__ ) self.assertEqual(old_configuration.hidden_size , 7_6_8 )
162
from __future__ import annotations from collections import deque from collections.abc import Sequence from dataclasses import dataclass from typing import Any @dataclass class __lowercase : """simple docstring""" _UpperCAmelCase = 42 _UpperCAmelCase = None _UpperCAmelCase = None def a__ ( ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = Node(1 ) SCREAMING_SNAKE_CASE_ : Any = Node(2 ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = Node(3 ) SCREAMING_SNAKE_CASE_ : int = Node(4 ) SCREAMING_SNAKE_CASE_ : List[str] = Node(5 ) return tree def a__ ( A__ ): return [root.data, *preorder(root.left ), *preorder(root.right )] if root else [] def a__ ( A__ ): return postorder(root.left ) + postorder(root.right ) + [root.data] if root else [] def a__ ( A__ ): return [*inorder(root.left ), root.data, *inorder(root.right )] if root else [] def a__ ( A__ ): return (max(height(root.left ), height(root.right ) ) + 1) if root else 0 def a__ ( A__ ): SCREAMING_SNAKE_CASE_ : list[Any] = [] if root is None: return output SCREAMING_SNAKE_CASE_ : int = deque([root] ) while process_queue: SCREAMING_SNAKE_CASE_ : List[str] = process_queue.popleft() output.append(node.data ) if node.left: process_queue.append(node.left ) if node.right: process_queue.append(node.right ) return output def a__ ( A__, A__ ): SCREAMING_SNAKE_CASE_ : list[Any] = [] def populate_output(A__, A__ ) -> None: if not root: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.left, level - 1 ) populate_output(root.right, level - 1 ) populate_output(A__, A__ ) return output def a__ ( A__, A__ ): SCREAMING_SNAKE_CASE_ : list[Any] = [] def populate_output(A__, A__ ) -> None: if root is None: return if level == 1: output.append(root.data ) elif level > 1: populate_output(root.right, level - 1 ) populate_output(root.left, level - 1 ) populate_output(A__, A__ ) return output def a__ ( A__ ): if root is None: return [] SCREAMING_SNAKE_CASE_ : list[Sequence[Node | None]] = [] SCREAMING_SNAKE_CASE_ : Optional[Any] = 0 SCREAMING_SNAKE_CASE_ : Optional[Any] = height(A__ ) for h in range(1, height_tree + 1 ): if not flag: output.append(get_nodes_from_left_to_right(A__, A__ ) ) SCREAMING_SNAKE_CASE_ : List[Any] = 1 else: output.append(get_nodes_from_right_to_left(A__, A__ ) ) SCREAMING_SNAKE_CASE_ : Optional[Any] = 0 return output def a__ ( ): # Main function for testing. SCREAMING_SNAKE_CASE_ : Optional[int] = make_tree() print(F'''In-order Traversal: {inorder(A__ )}''' ) print(F'''Pre-order Traversal: {preorder(A__ )}''' ) print(F'''Post-order Traversal: {postorder(A__ )}''', '\n' ) print(F'''Height of Tree: {height(A__ )}''', '\n' ) print('Complete Level Order Traversal: ' ) print(level_order(A__ ), '\n' ) print('Level-wise order Traversal: ' ) for level in range(1, height(A__ ) + 1 ): print(F'''Level {level}:''', get_nodes_from_left_to_right(A__, level=A__ ) ) print('\nZigZag order Traversal: ' ) print(zigzag(A__ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
162
1
'''simple docstring''' import os import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from huggingface_hub.file_download import http_get from requests.exceptions import HTTPError from transformers import ( AlbertTokenizer, AutoTokenizer, BertTokenizer, BertTokenizerFast, GPTaTokenizerFast, is_tokenizers_available, ) from transformers.testing_utils import TOKEN, USER, is_staging_test, require_tokenizers from transformers.tokenization_utils import Trie sys.path.append(str(Path(__file__).parent.parent / "utils")) from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class __snake_case ( unittest.TestCase): """simple docstring""" def __lowercase ( self : Tuple ) -> List[Any]: lowerCAmelCase_ : Optional[int] = mock.Mock() lowerCAmelCase_ : str = 5_00 lowerCAmelCase_ : str = {} lowerCAmelCase_ : Tuple = HTTPError lowerCAmelCase_ : Tuple = {} # Download this model to make sure it's in the cache. lowerCAmelCase_ : List[str] = BertTokenizer.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("""requests.Session.request""" , return_value=lowerCamelCase__ ) as mock_head: lowerCAmelCase_ : Optional[int] = BertTokenizer.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) # This check we did call the fake head request mock_head.assert_called() @require_tokenizers def __lowercase ( self : Optional[Any] ) -> Optional[int]: lowerCAmelCase_ : List[Any] = mock.Mock() lowerCAmelCase_ : int = 5_00 lowerCAmelCase_ : Any = {} lowerCAmelCase_ : Dict = HTTPError lowerCAmelCase_ : int = {} # Download this model to make sure it's in the cache. lowerCAmelCase_ : List[str] = GPTaTokenizerFast.from_pretrained("""gpt2""" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("""requests.Session.request""" , return_value=lowerCamelCase__ ) as mock_head: lowerCAmelCase_ : List[str] = GPTaTokenizerFast.from_pretrained("""gpt2""" ) # This check we did call the fake head request mock_head.assert_called() def __lowercase ( self : Optional[int] ) -> Union[str, Any]: # This test is for deprecated behavior and can be removed in v5 try: lowerCAmelCase_ : Optional[Any] = tempfile.mktemp() with open(lowerCamelCase__ , """wb""" ) as f: http_get("""https://huggingface.co/albert-base-v1/resolve/main/spiece.model""" , lowerCamelCase__ ) lowerCAmelCase_ : str = AlbertTokenizer.from_pretrained(lowerCamelCase__ ) finally: os.remove(lowerCamelCase__ ) # Supporting this legacy load introduced a weird bug where the tokenizer would load local files if they are in # the current folder and have the right name. if os.path.isfile("""tokenizer.json""" ): # We skip the test if the user has a `tokenizer.json` in this folder to avoid deleting it. return try: with open("""tokenizer.json""" , """wb""" ) as f: http_get("""https://huggingface.co/hf-internal-testing/tiny-random-bert/blob/main/tokenizer.json""" , lowerCamelCase__ ) lowerCAmelCase_ : str = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ) # The tiny random BERT has a vocab size of 1024, tiny gpt2 as a vocab size of 1000 self.assertEqual(tokenizer.vocab_size , 10_00 ) # Tokenizer should depend on the remote checkpoint, not the local tokenizer.json file. finally: os.remove("""tokenizer.json""" ) def __lowercase ( self : Optional[Any] ) -> Optional[int]: lowerCAmelCase_ : List[Any] = AlbertTokenizer.from_pretrained("""https://huggingface.co/albert-base-v1/resolve/main/spiece.model""" ) @is_staging_test class __snake_case ( unittest.TestCase): """simple docstring""" lowercase = ['[UNK]', '[CLS]', '[SEP]', '[PAD]', '[MASK]', 'bla', 'blou'] @classmethod def __lowercase ( cls : Union[str, Any] ) -> Tuple: lowerCAmelCase_ : Tuple = TOKEN HfFolder.save_token(lowerCamelCase__ ) @classmethod def __lowercase ( cls : str ) -> str: try: delete_repo(token=cls._token , repo_id="""test-tokenizer""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""valid_org/test-tokenizer-org""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""test-dynamic-tokenizer""" ) except HTTPError: pass def __lowercase ( self : int ) -> Any: with tempfile.TemporaryDirectory() as tmp_dir: lowerCAmelCase_ : Tuple = os.path.join(lowerCamelCase__ , """vocab.txt""" ) with open(lowerCamelCase__ , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in self.vocab_tokens] ) ) lowerCAmelCase_ : List[str] = BertTokenizer(lowerCamelCase__ ) tokenizer.push_to_hub("""test-tokenizer""" , use_auth_token=self._token ) lowerCAmelCase_ : List[Any] = BertTokenizer.from_pretrained(F'{USER}/test-tokenizer' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id="""test-tokenizer""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(lowerCamelCase__ , repo_id="""test-tokenizer""" , push_to_hub=lowerCamelCase__ , use_auth_token=self._token ) lowerCAmelCase_ : int = BertTokenizer.from_pretrained(F'{USER}/test-tokenizer' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) def __lowercase ( self : Any ) -> int: with tempfile.TemporaryDirectory() as tmp_dir: lowerCAmelCase_ : Tuple = os.path.join(lowerCamelCase__ , """vocab.txt""" ) with open(lowerCamelCase__ , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in self.vocab_tokens] ) ) lowerCAmelCase_ : List[str] = BertTokenizer(lowerCamelCase__ ) tokenizer.push_to_hub("""valid_org/test-tokenizer-org""" , use_auth_token=self._token ) lowerCAmelCase_ : str = BertTokenizer.from_pretrained("""valid_org/test-tokenizer-org""" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id="""valid_org/test-tokenizer-org""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained( lowerCamelCase__ , repo_id="""valid_org/test-tokenizer-org""" , push_to_hub=lowerCamelCase__ , use_auth_token=self._token ) lowerCAmelCase_ : Dict = BertTokenizer.from_pretrained("""valid_org/test-tokenizer-org""" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) @require_tokenizers def __lowercase ( self : Union[str, Any] ) -> Optional[int]: CustomTokenizer.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: lowerCAmelCase_ : str = os.path.join(lowerCamelCase__ , """vocab.txt""" ) with open(lowerCamelCase__ , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in self.vocab_tokens] ) ) lowerCAmelCase_ : List[Any] = CustomTokenizer(lowerCamelCase__ ) # No fast custom tokenizer tokenizer.push_to_hub("""test-dynamic-tokenizer""" , use_auth_token=self._token ) lowerCAmelCase_ : Tuple = AutoTokenizer.from_pretrained(F'{USER}/test-dynamic-tokenizer' , trust_remote_code=lowerCamelCase__ ) # Can't make an isinstance check because the new_model.config is from the CustomTokenizer class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , """CustomTokenizer""" ) # Fast and slow custom tokenizer CustomTokenizerFast.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: lowerCAmelCase_ : Optional[int] = os.path.join(lowerCamelCase__ , """vocab.txt""" ) with open(lowerCamelCase__ , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in self.vocab_tokens] ) ) lowerCAmelCase_ : Union[str, Any] = BertTokenizerFast.from_pretrained(lowerCamelCase__ ) bert_tokenizer.save_pretrained(lowerCamelCase__ ) lowerCAmelCase_ : Optional[int] = CustomTokenizerFast.from_pretrained(lowerCamelCase__ ) tokenizer.push_to_hub("""test-dynamic-tokenizer""" , use_auth_token=self._token ) lowerCAmelCase_ : Union[str, Any] = AutoTokenizer.from_pretrained(F'{USER}/test-dynamic-tokenizer' , trust_remote_code=lowerCamelCase__ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , """CustomTokenizerFast""" ) lowerCAmelCase_ : Dict = AutoTokenizer.from_pretrained( F'{USER}/test-dynamic-tokenizer' , use_fast=lowerCamelCase__ , trust_remote_code=lowerCamelCase__ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , """CustomTokenizer""" ) class __snake_case ( unittest.TestCase): """simple docstring""" def __lowercase ( self : Any ) -> Any: lowerCAmelCase_ : Optional[Any] = Trie() trie.add("""Hello 友達""" ) self.assertEqual(trie.data , {"""H""": {"""e""": {"""l""": {"""l""": {"""o""": {""" """: {"""友""": {"""達""": {"""""": 1}}}}}}}}} ) trie.add("""Hello""" ) trie.data self.assertEqual(trie.data , {"""H""": {"""e""": {"""l""": {"""l""": {"""o""": {"""""": 1, """ """: {"""友""": {"""達""": {"""""": 1}}}}}}}}} ) def __lowercase ( self : List[Any] ) -> Any: lowerCAmelCase_ : Union[str, Any] = Trie() self.assertEqual(trie.split("""[CLS] This is a extra_id_100""" ) , ["""[CLS] This is a extra_id_100"""] ) trie.add("""[CLS]""" ) trie.add("""extra_id_1""" ) trie.add("""extra_id_100""" ) self.assertEqual(trie.split("""[CLS] This is a extra_id_100""" ) , ["""[CLS]""", """ This is a """, """extra_id_100"""] ) def __lowercase ( self : List[Any] ) -> Dict: lowerCAmelCase_ : str = Trie() trie.add("""A""" ) self.assertEqual(trie.split("""ABC""" ) , ["""A""", """BC"""] ) self.assertEqual(trie.split("""BCA""" ) , ["""BC""", """A"""] ) def __lowercase ( self : Tuple ) -> Tuple: lowerCAmelCase_ : int = Trie() trie.add("""TOKEN]""" ) trie.add("""[SPECIAL_TOKEN]""" ) self.assertEqual(trie.split("""This is something [SPECIAL_TOKEN]""" ) , ["""This is something """, """[SPECIAL_TOKEN]"""] ) def __lowercase ( self : Tuple ) -> Any: lowerCAmelCase_ : Optional[int] = Trie() trie.add("""A""" ) trie.add("""P""" ) trie.add("""[SPECIAL_TOKEN]""" ) self.assertEqual(trie.split("""This is something [SPECIAL_TOKEN]""" ) , ["""This is something """, """[SPECIAL_TOKEN]"""] ) def __lowercase ( self : Optional[Any] ) -> Optional[int]: lowerCAmelCase_ : List[Any] = Trie() trie.add("""AB""" ) trie.add("""B""" ) trie.add("""C""" ) self.assertEqual(trie.split("""ABC""" ) , ["""AB""", """C"""] ) def __lowercase ( self : int ) -> int: lowerCAmelCase_ : Any = Trie() trie.add("""ABC""" ) trie.add("""B""" ) trie.add("""CD""" ) self.assertEqual(trie.split("""ABCD""" ) , ["""ABC""", """D"""] ) def __lowercase ( self : Union[str, Any] ) -> Optional[Any]: lowerCAmelCase_ : Dict = Trie() lowerCAmelCase_ : Optional[Any] = trie.cut_text("""ABC""" , [0, 0, 2, 1, 2, 3] ) self.assertEqual(lowerCamelCase__ , ["""AB""", """C"""] )
120
'''simple docstring''' from __future__ import annotations from collections.abc import MutableSequence class lowercase__ : def __init__( self : List[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : MutableSequence[float] ): '''simple docstring''' if len(lowerCamelCase__ ) != degree + 1: raise ValueError( 'The number of coefficients should be equal to the degree + 1.' ) _UpperCamelCase : list[float] = list(lowerCamelCase__ ) _UpperCamelCase : Tuple = degree def __add__( self : Optional[int] ,lowerCamelCase__ : Polynomial ): '''simple docstring''' if self.degree > polynomial_a.degree: _UpperCamelCase : str = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree ,lowerCamelCase__ ) else: _UpperCamelCase : str = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree ,lowerCamelCase__ ) def __sub__( self : Dict ,lowerCamelCase__ : Polynomial ): '''simple docstring''' return self + polynomial_a * Polynomial(0 ,[-1] ) def __neg__( self : Dict ): '''simple docstring''' return Polynomial(self.degree ,[-c for c in self.coefficients] ) def __mul__( self : Union[str, Any] ,lowerCamelCase__ : Polynomial ): '''simple docstring''' _UpperCamelCase : list[float] = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree ,lowerCamelCase__ ) def UpperCamelCase_ ( self : Dict ,lowerCamelCase__ : int | float ): '''simple docstring''' _UpperCamelCase : int | float = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : Union[str, Any] ): '''simple docstring''' _UpperCamelCase : Dict = '' for i in range(self.degree ,-1 ,-1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(lowerCamelCase__ ) return polynomial def __repr__( self : List[str] ): '''simple docstring''' return self.__str__() def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : list[float] = [0] * self.degree for i in range(self.degree ): _UpperCamelCase : Optional[int] = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 ,lowerCamelCase__ ) def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : int | float = 0 ): '''simple docstring''' _UpperCamelCase : list[float] = [0] * (self.degree + 2) _UpperCamelCase : Any = constant for i in range(self.degree + 1 ): _UpperCamelCase : Optional[Any] = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 ,lowerCamelCase__ ) def __eq__( self : str ,lowerCamelCase__ : object ): '''simple docstring''' if not isinstance(lowerCamelCase__ ,lowerCamelCase__ ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : List[str] ,lowerCamelCase__ : object ): '''simple docstring''' return not self.__eq__(lowerCamelCase__ )
83
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase : List[str] = logging.get_logger(__name__) lowerCAmelCase : Union[str, Any] = { "andreasmadsen/efficient_mlm_m0.40": ( "https://huggingface.co/andreasmadsen/efficient_mlm_m0.40/resolve/main/config.json" ), } class SCREAMING_SNAKE_CASE__ ( snake_case_): lowerCAmelCase_ = """roberta-prelayernorm""" def __init__( self , A_=50265 , A_=768 , A_=12 , A_=12 , A_=3072 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=2 , A_=0.02 , A_=1e-12 , A_=1 , A_=0 , A_=2 , A_="absolute" , A_=True , A_=None , **A_ , )-> Dict: '''simple docstring''' super().__init__(pad_token_id=lowerCamelCase_ , bos_token_id=lowerCamelCase_ , eos_token_id=lowerCamelCase_ , **lowerCamelCase_ ) UpperCamelCase = vocab_size UpperCamelCase = hidden_size UpperCamelCase = num_hidden_layers UpperCamelCase = num_attention_heads UpperCamelCase = hidden_act UpperCamelCase = intermediate_size UpperCamelCase = hidden_dropout_prob UpperCamelCase = attention_probs_dropout_prob UpperCamelCase = max_position_embeddings UpperCamelCase = type_vocab_size UpperCamelCase = initializer_range UpperCamelCase = layer_norm_eps UpperCamelCase = position_embedding_type UpperCamelCase = use_cache UpperCamelCase = classifier_dropout class SCREAMING_SNAKE_CASE__ ( snake_case_): @property def UpperCAmelCase_ ( self )-> Tuple: '''simple docstring''' if self.task == "multiple-choice": UpperCamelCase = {0: 'batch', 1: 'choice', 2: 'sequence'} else: UpperCamelCase = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
355
'''simple docstring''' import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class SCREAMING_SNAKE_CASE__ ( snake_case_ , snake_case_ , snake_case_): @register_to_config def __init__( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ , A_ , A_ , A_ = False , )-> Optional[Any]: '''simple docstring''' super().__init__() UpperCamelCase = nn.Embedding(A_ , A_ ) UpperCamelCase = nn.Embedding(A_ , A_ ) UpperCamelCase = False UpperCamelCase = nn.Dropout(p=A_ ) UpperCamelCase = TaConfig( vocab_size=A_ , d_model=A_ , num_heads=A_ , d_kv=A_ , d_ff=A_ , dropout_rate=A_ , feed_forward_proj=A_ , is_decoder=A_ , is_encoder_decoder=A_ , ) UpperCamelCase = nn.ModuleList() for lyr_num in range(A_ ): UpperCamelCase = TaBlock(A_ ) self.encoders.append(A_ ) UpperCamelCase = TaLayerNorm(A_ ) UpperCamelCase = nn.Dropout(p=A_ ) def UpperCAmelCase_ ( self , A_ , A_ )-> Tuple: '''simple docstring''' UpperCamelCase = self.token_embedder(A_ ) UpperCamelCase = encoder_input_tokens.shape[1] UpperCamelCase = torch.arange(A_ , device=encoder_input_tokens.device ) x += self.position_encoding(A_ ) UpperCamelCase = self.dropout_pre(A_ ) # inverted the attention mask UpperCamelCase = encoder_input_tokens.size() UpperCamelCase = self.get_extended_attention_mask(A_ , A_ ) for lyr in self.encoders: UpperCamelCase = lyr(A_ , A_ )[0] UpperCamelCase = self.layer_norm(A_ ) return self.dropout_post(A_ ), encoder_inputs_mask
251
0
'''simple docstring''' from math import factorial lowerCamelCase_ = {str(digit): factorial(digit) for digit in range(10)} def __lowercase ( __lowercase ) -> int: '''simple docstring''' if not isinstance(__lowercase , __lowercase ): raise TypeError("Parameter number must be int" ) if number < 0: raise ValueError("Parameter number must be greater than or equal to 0" ) # Converts number in string to iterate on its digits and adds its factorial. return sum(DIGIT_FACTORIAL[digit] for digit in str(__lowercase ) ) def __lowercase ( __lowercase = 60 , __lowercase = 100_0000 ) -> int: '''simple docstring''' if not isinstance(__lowercase , __lowercase ) or not isinstance(__lowercase , __lowercase ): raise TypeError("Parameters chain_length and number_limit must be int" ) if chain_length <= 0 or number_limit <= 0: raise ValueError( "Parameters chain_length and number_limit must be greater than 0" ) # the counter for the chains with the exact desired length _A = 0 # the cached sizes of the previous chains _A = {} for start_chain_element in range(1 , __lowercase ): # The temporary set will contain the elements of the chain _A = set() _A = 0 # Stop computing the chain when you find a cached size, a repeating item or the # length is greater then the desired one. _A = start_chain_element while ( chain_element not in chain_sets_lengths and chain_element not in chain_set and chain_set_length <= chain_length ): chain_set.add(__lowercase ) chain_set_length += 1 _A = digit_factorial_sum(__lowercase ) if chain_element in chain_sets_lengths: chain_set_length += chain_sets_lengths[chain_element] _A = chain_set_length # If chain contains the exact amount of elements increase the counter if chain_set_length == chain_length: chains_counter += 1 return chains_counter if __name__ == "__main__": import doctest doctest.testmod() print(F"""{solution()}""")
79
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from argparse import ArgumentParser from accelerate.commands.config import get_config_parser from accelerate.commands.env import env_command_parser from accelerate.commands.launch import launch_command_parser from accelerate.commands.test import test_command_parser from accelerate.commands.tpu import tpu_command_parser def _SCREAMING_SNAKE_CASE ( ) ->Dict: '''simple docstring''' a : Any = ArgumentParser("Accelerate CLI tool" , usage="accelerate <command> [<args>]" , allow_abbrev=_lowercase ) a : List[Any] = parser.add_subparsers(help="accelerate command helpers" ) # Register commands get_config_parser(subparsers=_lowercase ) env_command_parser(subparsers=_lowercase ) launch_command_parser(subparsers=_lowercase ) tpu_command_parser(subparsers=_lowercase ) test_command_parser(subparsers=_lowercase ) # Let's go a : List[Any] = parser.parse_args() if not hasattr(_lowercase , "func" ): parser.print_help() exit(1 ) # Run args.func(_lowercase ) if __name__ == "__main__": main()
105
0
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, Features, Value from .base import TaskTemplate @dataclass(frozen=lowerCAmelCase__ ) class snake_case__ ( lowerCAmelCase__ ): lowercase__ : str = field(default='''automatic-speech-recognition''' , metadata={'''include_in_asdict_even_if_is_default''': True} ) lowercase__ : ClassVar[Features] = Features({'''audio''': Audio()} ) lowercase__ : ClassVar[Features] = Features({'''transcription''': Value('''string''' )} ) lowercase__ : str = "audio" lowercase__ : str = "transcription" def __magic_name__ ( self , lowerCAmelCase__ ) -> Union[str, Any]: if self.audio_column not in features: raise ValueError(F'Column {self.audio_column} is not present in features.' ) if not isinstance(features[self.audio_column] , _SCREAMING_SNAKE_CASE ): raise ValueError(F'Column {self.audio_column} is not an Audio type.' ) __magic_name__ : Union[str, Any] = copy.deepcopy(self ) __magic_name__ : int = self.input_schema.copy() __magic_name__ : Any = features[self.audio_column] __magic_name__ : List[str] = input_schema return task_template @property def __magic_name__ ( self ) -> Dict[str, str]: return {self.audio_column: "audio", self.transcription_column: "transcription"}
355
import collections import json import os import re from typing import TYPE_CHECKING, List, Optional, Tuple import numpy as np from ...tokenization_utils_fast import PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __magic_name__: int = logging.get_logger(__name__) __magic_name__: Any = {"vocab_file": "vocab.txt", "emoji_file": "emoji.json"} __magic_name__: Union[str, Any] = { "vocab_file": { "abeja/gpt-neox-japanese-2.7b": "https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/vocab.txt", }, "emoji_file": { "abeja/gpt-neox-japanese-2.7b": "https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/emoji.json", }, } __magic_name__: Optional[Any] = { "abeja/gpt-neox-japanese-2.7b": 2_048, } def UpperCamelCase ( _A, _A ): """simple docstring""" with open(_A, """r""", encoding="""utf-8""" ) as f: __magic_name__ : Tuple = json.loads(f.read() ) __magic_name__ : Optional[Any] = collections.OrderedDict() __magic_name__ : List[str] = collections.OrderedDict() __magic_name__ : int = collections.OrderedDict() with open(_A, """r""", encoding="""utf-8""" ) as f: __magic_name__ : Tuple = f.readlines() __magic_name__ : Optional[int] = [[t.rstrip("""\n""" )] if (t == """,""" or """,""" not in t) else t.rstrip("""\n""" ).split(""",""" ) for t in token] for idx, b in enumerate(_A ): __magic_name__ : Dict = b __magic_name__ : Optional[Any] = idx for wd in b: __magic_name__ : Dict = idx return vocab, raw_vocab, ids_to_tokens, emoji class snake_case__ ( _lowerCAmelCase ): lowercase__ : Union[str, Any] = VOCAB_FILES_NAMES lowercase__ : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP lowercase__ : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ : List[str] = ['''input_ids''', '''attention_mask'''] def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__="<|endoftext|>" , lowerCAmelCase__="<|endoftext|>" , lowerCAmelCase__="<|startoftext|>" , lowerCAmelCase__="<|endoftext|>" , lowerCAmelCase__=False , **lowerCAmelCase__ , ) -> str: super().__init__( unk_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , do_clean_text=lowerCAmelCase__ , **lowerCAmelCase__ , ) if not os.path.isfile(lowerCAmelCase__ ): raise ValueError( F'Can\'t find a vocabulary file at path \'{vocab_file}\'. To load the vocabulary from a Google pretrained' """ model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`""" ) if not os.path.isfile(lowerCAmelCase__ ): raise ValueError( F'Can\'t find a emoji file at path \'{emoji_file}\'. To load the emoji information from a Google' """ pretrained model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`""" ) __magic_name__ : Dict = do_clean_text __magic_name__ ,__magic_name__ ,__magic_name__ ,__magic_name__ : Union[str, Any] = load_vocab_and_emoji(lowerCAmelCase__ , lowerCAmelCase__ ) __magic_name__ : str = SubWordJapaneseTokenizer( vocab=self.vocab , ids_to_tokens=self.ids_to_tokens , emoji=self.emoji ) @property def __magic_name__ ( self ) -> Union[str, Any]: # self.vocab contains support for character fluctuation unique to Japanese, and has a large number of vocab return len(self.raw_vocab ) def __magic_name__ ( self ) -> List[Any]: return dict(self.raw_vocab , **self.added_tokens_encoder ) def __magic_name__ ( self , lowerCAmelCase__ ) -> Any: return self.subword_tokenizer.tokenize(lowerCAmelCase__ , clean=self.do_clean_text ) def __magic_name__ ( self , lowerCAmelCase__ ) -> int: return self.vocab.get(lowerCAmelCase__ , self.vocab.get(self.unk_token ) ) def __magic_name__ ( self , lowerCAmelCase__ ) -> Dict: return self.subword_tokenizer.convert_id_to_token(lowerCAmelCase__ ) def __magic_name__ ( self , lowerCAmelCase__ ) -> Any: __magic_name__ : int = """""".join(lowerCAmelCase__ ).strip() return out_string def __magic_name__ ( self , lowerCAmelCase__ ) -> List[int]: __magic_name__ : Optional[Any] = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) + [self.eos_token_id] ) if len(lowerCAmelCase__ ) > self.model_max_length: __magic_name__ : int = input_ids[-self.model_max_length :] return input_ids def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: __magic_name__ : Dict = 0 if os.path.isdir(lowerCAmelCase__ ): __magic_name__ : Union[str, Any] = os.path.join( lowerCAmelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) __magic_name__ : Optional[Any] = os.path.join( lowerCAmelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""emoji_file"""] ) else: __magic_name__ : Optional[Any] = ( (filename_prefix + """-""" if filename_prefix else """""") + save_directory + VOCAB_FILES_NAMES["""vocab_file"""] ) __magic_name__ : Any = ( (filename_prefix + """-""" if filename_prefix else """""") + save_directory + VOCAB_FILES_NAMES["""emoji_file"""] ) with open(lowerCAmelCase__ , """w""" , encoding="""utf-8""" ) as writer: for token_index, token in self.ids_to_tokens.items(): if index != token_index: logger.warning( F'Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.' """ Please check that the vocabulary is not corrupted!""" ) __magic_name__ : Optional[Any] = token_index writer.write(""",""".join(lowerCAmelCase__ ) + """\n""" ) index += 1 with open(lowerCAmelCase__ , """w""" , encoding="""utf-8""" ) as writer: json.dump(self.emoji , lowerCAmelCase__ ) return vocab_file, emoji_file class snake_case__ ( _lowerCAmelCase ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> str: __magic_name__ : str = vocab # same as swe __magic_name__ : int = ids_to_tokens # same as bpe __magic_name__ : List[str] = emoji __magic_name__ : Optional[Any] = np.max([len(lowerCAmelCase__ ) for w in self.vocab.keys()] ) __magic_name__ : str = re.compile(R"""(https?|ftp)(:\/\/[-_\.!~*\'()a-zA-Z0-9;\/?:\@&=\+$,%#]+)""" ) __magic_name__ : str = re.compile(R"""[A-Za-z0-9\._+]*@[\-_0-9A-Za-z]+(\.[A-Za-z]+)*""" ) __magic_name__ : Any = re.compile(R"""[\(]{0,1}[0-9]{2,4}[\)\-\(]{0,1}[0-9]{2,4}[\)\-]{0,1}[0-9]{3,4}""" ) __magic_name__ : List[str] = re.compile( R"""([12]\d{3}[/\-年])*(0?[1-9]|1[0-2])[/\-月]((0?[1-9]|[12][0-9]|3[01])日?)*(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*""" ) __magic_name__ : Dict = re.compile( R"""(明治|大正|昭和|平成|令和|㍾|㍽|㍼|㍻|\u32ff)\d{1,2}年(0?[1-9]|1[0-2])月(0?[1-9]|[12][0-9]|3[01])日(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*""" ) __magic_name__ : List[Any] = re.compile( R"""((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*億)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*万)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*千)*(0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*(千円|万円|千万円|円|千ドル|万ドル|千万ドル|ドル|千ユーロ|万ユーロ|千万ユーロ|ユーロ)+(\(税込\)|\(税抜\)|\+tax)*""" ) __magic_name__ : str = """─━│┃┄┅┆┇┈┉┊┋┌┍┎┏┐┑┒┓└┕┖┗┘┙┚┛├┝┞┟┠┡┢┣┤┥┦┧┨┩┪┫┬┭┮┯┰┱┲┳┴┵┶┷┸┹┺┻┼┽┾┿╀╁╂╃╄╅╆╇╈╉╊╋╌╍╎╏═║╒╓╔╕╖╗╘╙╚╛╜╝╞╟╠╡╢╣╤╥╦╧╨╩╪╫╬╭╮╯╰╱╲╳╴╵╶╷╸╹╺╻╼╽╾╿""" __magic_name__ : Dict = """▀▁▂▃▄▅▆▇█▉▊▋▌▍▎▏▐░▒▓▔▕▖▗▘▙▚▛▜▝▞▟""" __magic_name__ : Optional[int] = str.maketrans({k: """<BLOCK>""" for k in keisen + blocks} ) def __len__( self ) -> int: return len(self.ids_to_tokens ) def __magic_name__ ( self , lowerCAmelCase__ ) -> Optional[int]: __magic_name__ : Any = self.content_repattera.sub("""<URL>""" , lowerCAmelCase__ ) __magic_name__ : Any = self.content_repattera.sub("""<EMAIL>""" , lowerCAmelCase__ ) __magic_name__ : Dict = self.content_repattera.sub("""<TEL>""" , lowerCAmelCase__ ) __magic_name__ : Tuple = self.content_repattera.sub("""<DATE>""" , lowerCAmelCase__ ) __magic_name__ : str = self.content_repattera.sub("""<DATE>""" , lowerCAmelCase__ ) __magic_name__ : Tuple = self.content_repattera.sub("""<PRICE>""" , lowerCAmelCase__ ) __magic_name__ : Dict = content.translate(self.content_transa ) while "<BLOCK><BLOCK>" in content: __magic_name__ : str = content.replace("""<BLOCK><BLOCK>""" , """<BLOCK>""" ) return content def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__=False ) -> Any: __magic_name__ : Union[str, Any] = text.replace(""" """ , """<SP>""" ) __magic_name__ : Optional[int] = text.replace(""" """ , """<SP>""" ) __magic_name__ : int = text.replace("""\r\n""" , """<BR>""" ) __magic_name__ : Optional[Any] = text.replace("""\n""" , """<BR>""" ) __magic_name__ : List[Any] = text.replace("""\r""" , """<BR>""" ) __magic_name__ : List[str] = text.replace("""\t""" , """<TAB>""" ) __magic_name__ : Optional[Any] = text.replace("""—""" , """ー""" ) __magic_name__ : int = text.replace("""−""" , """ー""" ) for k, v in self.emoji["emoji"].items(): if k in text: __magic_name__ : List[Any] = text.replace(lowerCAmelCase__ , lowerCAmelCase__ ) if clean: __magic_name__ : List[str] = self.clean_text(lowerCAmelCase__ ) def check_simbol(lowerCAmelCase__ ): __magic_name__ : int = x.encode() if len(lowerCAmelCase__ ) == 1 and len(lowerCAmelCase__ ) == 2: __magic_name__ : str = (int(e[0] ) << 8) + int(e[1] ) if ( (c >= 0Xc2_a1 and c <= 0Xc2_bf) or (c >= 0Xc7_80 and c <= 0Xc7_83) or (c >= 0Xca_b9 and c <= 0Xcb_bf) or (c >= 0Xcc_80 and c <= 0Xcd_a2) ): return True return False def checkuae(lowerCAmelCase__ ): __magic_name__ : List[Any] = x.encode() if len(lowerCAmelCase__ ) == 1 and len(lowerCAmelCase__ ) == 3: __magic_name__ : Dict = (int(e[0] ) << 16) + (int(e[1] ) << 8) + int(e[2] ) if c >= 0Xe2_80_80 and c <= 0Xe2_b0_7f: return True return False __magic_name__ : Optional[Any] = 0 __magic_name__ : List[str] = [] while pos < len(lowerCAmelCase__ ): __magic_name__ : Optional[int] = min(len(lowerCAmelCase__ ) , pos + self.maxlen + 1 ) if text[pos] == """<""" else pos + 3 __magic_name__ : int = [] # (token_id, token, pos) for e in range(lowerCAmelCase__ , lowerCAmelCase__ , -1 ): __magic_name__ : str = text[pos:e] if wd in self.vocab: if wd[0] == "<" and len(lowerCAmelCase__ ) > 2: __magic_name__ : Union[str, Any] = [(self.vocab[wd], wd, e)] break else: candidates.append((self.vocab[wd], wd, e) ) if len(lowerCAmelCase__ ) > 0: # the smallest token_id is adopted __magic_name__ ,__magic_name__ ,__magic_name__ : str = sorted(lowerCAmelCase__ , key=lambda lowerCAmelCase__ : x[0] )[0] result.append(lowerCAmelCase__ ) __magic_name__ : List[Any] = e else: __magic_name__ : Dict = pos + 1 __magic_name__ : str = text[pos:end] if check_simbol(lowerCAmelCase__ ): result.append("""<KIGOU>""" ) elif checkuae(lowerCAmelCase__ ): result.append("""<U2000U2BFF>""" ) else: for i in wd.encode("""utf-8""" ): result.append("""<|byte%d|>""" % i ) __magic_name__ : List[str] = end return result def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__="\n" ) -> Optional[int]: __magic_name__ : Tuple = [] __magic_name__ : Tuple = [] __magic_name__ : str = self.ids_to_tokens[index][0] if word[:6] == "<|byte" and word[-2:] == "|>": byte_tokens.append(int(word[6:-2] ) ) else: if len(lowerCAmelCase__ ) > 0: words.append(bytearray(lowerCAmelCase__ ).decode("""utf-8""" , errors="""replace""" ) ) __magic_name__ : List[Any] = [] if word[:7] == "<|emoji" and word[-2:] == "|>": words.append(self.emoji["""emoji_inv"""][word] ) elif word == "<SP>": words.append(""" """ ) elif word == "<BR>": words.append(lowerCAmelCase__ ) elif word == "<TAB>": words.append("""\t""" ) elif word == "<BLOCK>": words.append("""▀""" ) elif word == "<KIGOU>": words.append("""ǀ""" ) elif word == "<U2000U2BFF>": words.append("""‖""" ) else: words.append(lowerCAmelCase__ ) if len(lowerCAmelCase__ ) > 0: words.append(bytearray(lowerCAmelCase__ ).decode("""utf-8""" , errors="""replace""" ) ) __magic_name__ : Union[str, Any] = """""".join(lowerCAmelCase__ ) return text
138
0
"""simple docstring""" from PIL import Image def snake_case_ ( A_ : Image, A_ : float ): '''simple docstring''' def brightness(A_ : int ) -> float: return 1_28 + level + (c - 1_28) if not -255.0 <= level <= 255.0: raise ValueError('''level must be between -255.0 (black) and 255.0 (white)''' ) return img.point(A_ ) if __name__ == "__main__": # Load image with Image.open('''image_data/lena.jpg''') as img: # Change brightness to 100 lowerCAmelCase__ = change_brightness(img, 100) brigt_img.save('''image_data/lena_brightness.png''', format='''png''')
72
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase__ : Dict = logging.get_logger(__name__) UpperCAmelCase__ : str = { """facebook/nllb-moe-54B""": """https://huggingface.co/facebook/nllb-moe-54b/resolve/main/config.json""", } class a__ ( UpperCAmelCase ): """simple docstring""" UpperCAmelCase__ : List[Any] ="""nllb-moe""" UpperCAmelCase__ : Any =["""past_key_values"""] UpperCAmelCase__ : Dict ={"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : Union[str, Any] , UpperCAmelCase__ : List[Any]=1_2_8_1_1_2 , UpperCAmelCase__ : Tuple=1_0_2_4 , UpperCAmelCase__ : str=1_2 , UpperCAmelCase__ : int=4_0_9_6 , UpperCAmelCase__ : Dict=1_6 , UpperCAmelCase__ : Union[str, Any]=1_2 , UpperCAmelCase__ : int=4_0_9_6 , UpperCAmelCase__ : Optional[Any]=1_6 , UpperCAmelCase__ : Union[str, Any]=0.05 , UpperCAmelCase__ : Any=0.05 , UpperCAmelCase__ : List[Any]=True , UpperCAmelCase__ : Optional[Any]=True , UpperCAmelCase__ : Union[str, Any]="relu" , UpperCAmelCase__ : Dict=1_0_2_4 , UpperCAmelCase__ : Any=0.1 , UpperCAmelCase__ : str=0.1 , UpperCAmelCase__ : List[Any]=0.0 , UpperCAmelCase__ : Tuple=0.02 , UpperCAmelCase__ : Optional[int]=2 , UpperCAmelCase__ : str=True , UpperCAmelCase__ : Union[str, Any]=False , UpperCAmelCase__ : Dict="float32" , UpperCAmelCase__ : Any=False , UpperCAmelCase__ : Union[str, Any]=1_2_8 , UpperCAmelCase__ : Any=6_4 , UpperCAmelCase__ : Dict=4 , UpperCAmelCase__ : List[Any]=4 , UpperCAmelCase__ : Optional[Any]=0.0_01 , UpperCAmelCase__ : Optional[Any]=0.0_01 , UpperCAmelCase__ : Dict="all" , UpperCAmelCase__ : Optional[int]=False , UpperCAmelCase__ : List[Any]=False , UpperCAmelCase__ : List[str]=1.0 , UpperCAmelCase__ : Optional[int]=0.2 , UpperCAmelCase__ : Dict=1 , UpperCAmelCase__ : List[Any]=0 , UpperCAmelCase__ : Any=2 , UpperCAmelCase__ : Tuple=False , **UpperCAmelCase__ : Union[str, Any] , ) ->int: """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = vocab_size SCREAMING_SNAKE_CASE : Optional[int] = max_position_embeddings SCREAMING_SNAKE_CASE : str = d_model SCREAMING_SNAKE_CASE : Dict = encoder_ffn_dim SCREAMING_SNAKE_CASE : Union[str, Any] = encoder_layers SCREAMING_SNAKE_CASE : str = encoder_attention_heads SCREAMING_SNAKE_CASE : List[str] = decoder_ffn_dim SCREAMING_SNAKE_CASE : str = decoder_layers SCREAMING_SNAKE_CASE : Optional[Any] = decoder_attention_heads SCREAMING_SNAKE_CASE : Optional[Any] = dropout SCREAMING_SNAKE_CASE : Optional[int] = attention_dropout SCREAMING_SNAKE_CASE : str = activation_dropout SCREAMING_SNAKE_CASE : Dict = activation_function SCREAMING_SNAKE_CASE : str = init_std SCREAMING_SNAKE_CASE : Tuple = encoder_layerdrop SCREAMING_SNAKE_CASE : Tuple = decoder_layerdrop SCREAMING_SNAKE_CASE : List[str] = use_cache SCREAMING_SNAKE_CASE : Optional[Any] = encoder_layers SCREAMING_SNAKE_CASE : List[Any] = scale_embedding # scale factor will be sqrt(d_model) if True SCREAMING_SNAKE_CASE : Tuple = router_z_loss_coef SCREAMING_SNAKE_CASE : Tuple = router_aux_loss_coef SCREAMING_SNAKE_CASE : List[Any] = decoder_sparse_step SCREAMING_SNAKE_CASE : Any = encoder_sparse_step SCREAMING_SNAKE_CASE : Tuple = num_experts SCREAMING_SNAKE_CASE : Optional[int] = expert_capacity SCREAMING_SNAKE_CASE : int = router_bias if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(f"`router_dtype` must be one of 'float32', 'float16' or 'bfloat16', got {router_dtype}" ) SCREAMING_SNAKE_CASE : Optional[int] = router_dtype SCREAMING_SNAKE_CASE : Any = router_ignore_padding_tokens SCREAMING_SNAKE_CASE : Any = batch_prioritized_routing SCREAMING_SNAKE_CASE : Optional[Any] = second_expert_policy SCREAMING_SNAKE_CASE : Any = normalize_router_prob_before_dropping SCREAMING_SNAKE_CASE : Tuple = moe_eval_capacity_token_fraction SCREAMING_SNAKE_CASE : int = moe_token_dropout SCREAMING_SNAKE_CASE : Optional[int] = output_router_logits super().__init__( pad_token_id=UpperCAmelCase__ , bos_token_id=UpperCAmelCase__ , eos_token_id=UpperCAmelCase__ , is_encoder_decoder=UpperCAmelCase__ , decoder_start_token_id=UpperCAmelCase__ , **UpperCAmelCase__ , )
245
0
import argparse from transformers import BigBirdConfig, BigBirdForPreTraining, BigBirdForQuestionAnswering, load_tf_weights_in_big_bird from transformers.utils import logging logging.set_verbosity_info() def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' snake_case_ = BigBirdConfig.from_json_file(UpperCamelCase__ ) print(F'''Building PyTorch model from configuration: {config}''' ) if is_trivia_qa: snake_case_ = BigBirdForQuestionAnswering(UpperCamelCase__ ) else: snake_case_ = BigBirdForPreTraining(UpperCamelCase__ ) # Load weights from tf checkpoint load_tf_weights_in_big_bird(UpperCamelCase__ , UpperCamelCase__ , is_trivia_qa=UpperCamelCase__ ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(UpperCamelCase__ ) if __name__ == "__main__": _UpperCAmelCase : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--big_bird_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained BERT model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--is_trivia_qa""", action="""store_true""", help="""Whether to convert a model with a trivia_qa head.""" ) _UpperCAmelCase : str = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.tf_checkpoint_path, args.big_bird_config_file, args.pytorch_dump_path, args.is_trivia_qa )
200
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _UpperCAmelCase : Dict = { """configuration_blip""": [ """BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BlipConfig""", """BlipTextConfig""", """BlipVisionConfig""", ], """processing_blip""": ["""BlipProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Optional[int] = ["""BlipImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Any = [ """BLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """BlipModel""", """BlipPreTrainedModel""", """BlipForConditionalGeneration""", """BlipForQuestionAnswering""", """BlipVisionModel""", """BlipTextModel""", """BlipForImageTextRetrieval""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Any = [ """TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFBlipModel""", """TFBlipPreTrainedModel""", """TFBlipForConditionalGeneration""", """TFBlipForQuestionAnswering""", """TFBlipVisionModel""", """TFBlipTextModel""", """TFBlipForImageTextRetrieval""", ] if TYPE_CHECKING: from .configuration_blip import BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipConfig, BlipTextConfig, BlipVisionConfig from .processing_blip import BlipProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_blip import BlipImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip import ( BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, BlipModel, BlipPreTrainedModel, BlipTextModel, BlipVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blip import ( TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFBlipForConditionalGeneration, TFBlipForImageTextRetrieval, TFBlipForQuestionAnswering, TFBlipModel, TFBlipPreTrainedModel, TFBlipTextModel, TFBlipVisionModel, ) else: import sys _UpperCAmelCase : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
200
1
"""simple docstring""" import cmath import math def lowercase ( _SCREAMING_SNAKE_CASE : float , _SCREAMING_SNAKE_CASE : float , _SCREAMING_SNAKE_CASE : float , _SCREAMING_SNAKE_CASE : float ): '''simple docstring''' _UpperCAmelCase = math.radians(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = math.radians(_SCREAMING_SNAKE_CASE ) # Convert voltage and current to rectangular form _UpperCAmelCase = cmath.rect(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) _UpperCAmelCase = cmath.rect(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Calculate apparent power return voltage_rect * current_rect if __name__ == "__main__": import doctest doctest.testmod()
260
"""simple docstring""" from pathlib import Path import numpy as np from PIL import Image def lowercase ( _SCREAMING_SNAKE_CASE : np.ndarray ): '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = rgb[:, :, 0], rgb[:, :, 1], rgb[:, :, 2] return 0.2989 * r + 0.5870 * g + 0.1140 * b def lowercase ( _SCREAMING_SNAKE_CASE : np.ndarray ): '''simple docstring''' return (gray > 127) & (gray <= 255) def lowercase ( _SCREAMING_SNAKE_CASE : np.ndarray , _SCREAMING_SNAKE_CASE : np.ndarray ): '''simple docstring''' _UpperCAmelCase = np.zeros_like(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = np.zeros( (image.shape[0] + kernel.shape[0] - 1, image.shape[1] + kernel.shape[1] - 1) ) # Copy image to padded image _UpperCAmelCase = image # Iterate over image & apply kernel for x in range(image.shape[1] ): for y in range(image.shape[0] ): _UpperCAmelCase = ( kernel * image_padded[y : y + kernel.shape[0], x : x + kernel.shape[1]] ).sum() _UpperCAmelCase = int(summation > 0 ) return output if __name__ == "__main__": # read original image __A : str = Path(__file__).resolve().parent / "image_data" / "lena.jpg" __A : str = np.array(Image.open(lena_path)) # kernel to be applied __A : List[Any] = np.array([[0, 1, 0], [1, 1, 1], [0, 1, 0]]) __A : Optional[Any] = dilation(gray_to_binary(rgb_to_gray(lena)), structuring_element) # Save the output image __A : Optional[Any] = Image.fromarray(output).convert("RGB") pil_img.save("result_dilation.png")
260
1
"""simple docstring""" import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import ( AutoProcessor, BertTokenizerFast, BlipImageProcessor, GPTaTokenizer, InstructBlipProcessor, PreTrainedTokenizerFast, ) @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Tuple = tempfile.mkdtemp() lowerCAmelCase__ :Tuple = BlipImageProcessor() lowerCAmelCase__ :str = GPTaTokenizer.from_pretrained('hf-internal-testing/tiny-random-GPT2Model' ) lowerCAmelCase__ :str = BertTokenizerFast.from_pretrained('hf-internal-testing/tiny-random-bert' ) lowerCAmelCase__ :List[Any] = InstructBlipProcessor(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) processor.save_pretrained(self.tmpdirname ) def snake_case ( self , **__UpperCAmelCase ): '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **__UpperCAmelCase ).tokenizer def snake_case ( self , **__UpperCAmelCase ): '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **__UpperCAmelCase ).image_processor def snake_case ( self , **__UpperCAmelCase ): '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **__UpperCAmelCase ).qformer_tokenizer def snake_case ( self ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Tuple = [np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta )] lowerCAmelCase__ :Union[str, Any] = [Image.fromarray(np.moveaxis(__UpperCAmelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Tuple = InstructBlipProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() , qformer_tokenizer=self.get_qformer_tokenizer() , ) processor.save_pretrained(self.tmpdirname ) lowerCAmelCase__ :Any = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) lowerCAmelCase__ :List[str] = self.get_image_processor(do_normalize=__UpperCAmelCase , padding_value=1.0 ) lowerCAmelCase__ :Union[str, Any] = InstructBlipProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=__UpperCAmelCase , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , __UpperCAmelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __UpperCAmelCase ) self.assertIsInstance(processor.qformer_tokenizer , __UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :str = self.get_image_processor() lowerCAmelCase__ :List[str] = self.get_tokenizer() lowerCAmelCase__ :Any = self.get_qformer_tokenizer() lowerCAmelCase__ :Optional[Any] = InstructBlipProcessor( tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase , qformer_tokenizer=__UpperCAmelCase ) lowerCAmelCase__ :List[str] = self.prepare_image_inputs() lowerCAmelCase__ :Tuple = image_processor(__UpperCAmelCase , return_tensors='np' ) lowerCAmelCase__ :Optional[Any] = processor(images=__UpperCAmelCase , return_tensors='np' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :int = self.get_image_processor() lowerCAmelCase__ :Tuple = self.get_tokenizer() lowerCAmelCase__ :int = self.get_qformer_tokenizer() lowerCAmelCase__ :Any = InstructBlipProcessor( tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase , qformer_tokenizer=__UpperCAmelCase ) lowerCAmelCase__ :int = 'lower newer' lowerCAmelCase__ :int = processor(text=__UpperCAmelCase ) lowerCAmelCase__ :Optional[Any] = tokenizer(__UpperCAmelCase , return_token_type_ids=__UpperCAmelCase ) lowerCAmelCase__ :Dict = qformer_tokenizer(__UpperCAmelCase , return_token_type_ids=__UpperCAmelCase ) for key in encoded_tokens.keys(): self.assertListEqual(encoded_tokens[key] , encoded_processor[key] ) for key in encoded_tokens_qformer.keys(): self.assertListEqual(encoded_tokens_qformer[key] , encoded_processor['qformer_' + key] ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Tuple = self.get_image_processor() lowerCAmelCase__ :List[str] = self.get_tokenizer() lowerCAmelCase__ :Dict = self.get_qformer_tokenizer() lowerCAmelCase__ :Union[str, Any] = InstructBlipProcessor( tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase , qformer_tokenizer=__UpperCAmelCase ) lowerCAmelCase__ :int = 'lower newer' lowerCAmelCase__ :List[Any] = self.prepare_image_inputs() lowerCAmelCase__ :Any = processor(text=__UpperCAmelCase , images=__UpperCAmelCase ) self.assertListEqual( list(inputs.keys() ) , ['input_ids', 'attention_mask', 'qformer_input_ids', 'qformer_attention_mask', 'pixel_values'] , ) # test if it raises when no input is passed with pytest.raises(__UpperCAmelCase ): processor() def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :int = self.get_image_processor() lowerCAmelCase__ :int = self.get_tokenizer() lowerCAmelCase__ :List[str] = self.get_qformer_tokenizer() lowerCAmelCase__ :Optional[int] = InstructBlipProcessor( tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase , qformer_tokenizer=__UpperCAmelCase ) lowerCAmelCase__ :Optional[Any] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowerCAmelCase__ :List[str] = processor.batch_decode(__UpperCAmelCase ) lowerCAmelCase__ :Any = tokenizer.batch_decode(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :str = self.get_image_processor() lowerCAmelCase__ :List[str] = self.get_tokenizer() lowerCAmelCase__ :List[Any] = self.get_qformer_tokenizer() lowerCAmelCase__ :Optional[Any] = InstructBlipProcessor( tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase , qformer_tokenizer=__UpperCAmelCase ) lowerCAmelCase__ :Any = 'lower newer' lowerCAmelCase__ :Any = self.prepare_image_inputs() lowerCAmelCase__ :str = processor(text=__UpperCAmelCase , images=__UpperCAmelCase ) self.assertListEqual( list(inputs.keys() ) , ['input_ids', 'attention_mask', 'qformer_input_ids', 'qformer_attention_mask', 'pixel_values'] , )
254
"""simple docstring""" import inspect import os import re from transformers.configuration_utils import PretrainedConfig from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py __A = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. __A = direct_transformers_import(PATH_TO_TRANSFORMERS) __A = transformers.models.auto.configuration_auto.CONFIG_MAPPING __A = { # used to compute the property `self.chunk_length` """EncodecConfig""": ["""overlap"""], # used as `self.bert_model = BertModel(config, ...)` """DPRConfig""": True, # not used in modeling files, but it's an important information """FSMTConfig""": ["""langs"""], # used internally in the configuration class file """GPTNeoConfig""": ["""attention_types"""], # used internally in the configuration class file """EsmConfig""": ["""is_folding_model"""], # used during training (despite we don't have training script for these models yet) """Mask2FormerConfig""": ["""ignore_value"""], # `ignore_value` used during training (despite we don't have training script for these models yet) # `norm` used in conversion script (despite not using in the modeling file) """OneFormerConfig""": ["""ignore_value""", """norm"""], # used during preprocessing and collation, see `collating_graphormer.py` """GraphormerConfig""": ["""spatial_pos_max"""], # used internally in the configuration class file """T5Config""": ["""feed_forward_proj"""], # used internally in the configuration class file # `tokenizer_class` get default value `T5Tokenizer` intentionally """MT5Config""": ["""feed_forward_proj""", """tokenizer_class"""], """UMT5Config""": ["""feed_forward_proj""", """tokenizer_class"""], # used internally in the configuration class file """LongT5Config""": ["""feed_forward_proj"""], # used internally in the configuration class file """SwitchTransformersConfig""": ["""feed_forward_proj"""], # having default values other than `1e-5` - we can't fix them without breaking """BioGptConfig""": ["""layer_norm_eps"""], # having default values other than `1e-5` - we can't fix them without breaking """GLPNConfig""": ["""layer_norm_eps"""], # having default values other than `1e-5` - we can't fix them without breaking """SegformerConfig""": ["""layer_norm_eps"""], # having default values other than `1e-5` - we can't fix them without breaking """CvtConfig""": ["""layer_norm_eps"""], # having default values other than `1e-5` - we can't fix them without breaking """PerceiverConfig""": ["""layer_norm_eps"""], # used internally to calculate the feature size """InformerConfig""": ["""num_static_real_features""", """num_time_features"""], # used internally to calculate the feature size """TimeSeriesTransformerConfig""": ["""num_static_real_features""", """num_time_features"""], # used internally to calculate the feature size """AutoformerConfig""": ["""num_static_real_features""", """num_time_features"""], # used internally to calculate `mlp_dim` """SamVisionConfig""": ["""mlp_ratio"""], # For (head) training, but so far not implemented """ClapAudioConfig""": ["""num_classes"""], # Not used, but providing useful information to users """SpeechT5HifiGanConfig""": ["""sampling_rate"""], } # TODO (ydshieh): Check the failing cases, try to fix them or move some cases to the above block once we are sure SPECIAL_CASES_TO_ALLOW.update( { """CLIPSegConfig""": True, """DeformableDetrConfig""": True, """DetaConfig""": True, """DinatConfig""": True, """DonutSwinConfig""": True, """EfficientFormerConfig""": True, """FSMTConfig""": True, """JukeboxConfig""": True, """LayoutLMv2Config""": True, """MaskFormerSwinConfig""": True, """MT5Config""": True, """NatConfig""": True, """OneFormerConfig""": True, """PerceiverConfig""": True, """RagConfig""": True, """SpeechT5Config""": True, """SwinConfig""": True, """Swin2SRConfig""": True, """Swinv2Config""": True, """SwitchTransformersConfig""": True, """TableTransformerConfig""": True, """TapasConfig""": True, """TransfoXLConfig""": True, """UniSpeechConfig""": True, """UniSpeechSatConfig""": True, """WavLMConfig""": True, """WhisperConfig""": True, # TODO: @Arthur (for `alignment_head` and `alignment_layer`) """JukeboxPriorConfig""": True, # TODO: @Younes (for `is_decoder`) """Pix2StructTextConfig""": True, } ) def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->int: """simple docstring""" lowerCAmelCase__ :List[str] = False for attribute in attributes: for modeling_source in source_strings: # check if we can find `config.xxx`, `getattr(config, "xxx", ...)` or `getattr(self.config, "xxx", ...)` if ( F"config.{attribute}" in modeling_source or F"getattr(config, \"{attribute}\"" in modeling_source or F"getattr(self.config, \"{attribute}\"" in modeling_source ): lowerCAmelCase__ :List[str] = True # Deal with multi-line cases elif ( re.search( rF"getattr[ \t\v\n\r\f]*\([ \t\v\n\r\f]*(self\.)?config,[ \t\v\n\r\f]*\"{attribute}\"" , _SCREAMING_SNAKE_CASE , ) is not None ): lowerCAmelCase__ :int = True # `SequenceSummary` is called with `SequenceSummary(config)` elif attribute in [ "summary_type", "summary_use_proj", "summary_activation", "summary_last_dropout", "summary_proj_to_labels", "summary_first_dropout", ]: if "SequenceSummary" in modeling_source: lowerCAmelCase__ :Any = True if attribute_used: break if attribute_used: break # common and important attributes, even if they do not always appear in the modeling files lowerCAmelCase__ :Union[str, Any] = [ 'bos_index', 'eos_index', 'pad_index', 'unk_index', 'mask_index', 'image_size', 'use_cache', 'out_features', 'out_indices', ] lowerCAmelCase__ :Union[str, Any] = ['encoder_no_repeat_ngram_size'] # Special cases to be allowed lowerCAmelCase__ :Any = True if not attribute_used: lowerCAmelCase__ :List[Any] = False for attribute in attributes: # Allow if the default value in the configuration class is different from the one in `PretrainedConfig` if attribute in ["is_encoder_decoder"] and default_value is True: lowerCAmelCase__ :List[str] = True elif attribute in ["tie_word_embeddings"] and default_value is False: lowerCAmelCase__ :Tuple = True # Allow cases without checking the default value in the configuration class elif attribute in attributes_to_allow + attributes_used_in_generation: lowerCAmelCase__ :Optional[Any] = True elif attribute.endswith('_token_id' ): lowerCAmelCase__ :List[Any] = True # configuration class specific cases if not case_allowed: lowerCAmelCase__ :List[str] = SPECIAL_CASES_TO_ALLOW.get(config_class.__name__ , [] ) lowerCAmelCase__ :List[Any] = allowed_cases is True or attribute in allowed_cases return attribute_used or case_allowed def __A (_SCREAMING_SNAKE_CASE ) ->Union[str, Any]: """simple docstring""" lowerCAmelCase__ :List[Any] = dict(inspect.signature(config_class.__init__ ).parameters ) lowerCAmelCase__ :List[Any] = [x for x in list(signature.keys() ) if x not in ['self', 'kwargs']] lowerCAmelCase__ :List[Any] = [signature[param].default for param in parameter_names] # If `attribute_map` exists, an attribute can have different names to be used in the modeling files, and as long # as one variant is used, the test should pass lowerCAmelCase__ :Optional[Any] = {} if len(config_class.attribute_map ) > 0: lowerCAmelCase__ :Optional[int] = {v: k for k, v in config_class.attribute_map.items()} # Get the path to modeling source files lowerCAmelCase__ :str = inspect.getsourcefile(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :List[Any] = os.path.dirname(_SCREAMING_SNAKE_CASE ) # Let's check against all frameworks: as long as one framework uses an attribute, we are good. lowerCAmelCase__ :Dict = [os.path.join(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) for fn in os.listdir(_SCREAMING_SNAKE_CASE ) if fn.startswith('modeling_' )] # Get the source code strings lowerCAmelCase__ :Tuple = [] for path in modeling_paths: if os.path.isfile(_SCREAMING_SNAKE_CASE ): with open(_SCREAMING_SNAKE_CASE ) as fp: modeling_sources.append(fp.read() ) lowerCAmelCase__ :Any = [] for config_param, default_value in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): # `attributes` here is all the variant names for `config_param` lowerCAmelCase__ :Optional[int] = [config_param] # some configuration classes have non-empty `attribute_map`, and both names could be used in the # corresponding modeling files. As long as one of them appears, it is fine. if config_param in reversed_attribute_map: attributes.append(reversed_attribute_map[config_param] ) if not check_attribute_being_used(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): unused_attributes.append(attributes[0] ) return sorted(_SCREAMING_SNAKE_CASE ) def __A () ->List[Any]: """simple docstring""" lowerCAmelCase__ :Optional[int] = {} for _config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in _config_class.__module__: continue # Some config classes are not in `CONFIG_MAPPING` (e.g. `CLIPVisionConfig`, `Blip2VisionConfig`, etc.) lowerCAmelCase__ :List[str] = [ cls for name, cls in inspect.getmembers( inspect.getmodule(_config_class ) , lambda _SCREAMING_SNAKE_CASE : inspect.isclass(_SCREAMING_SNAKE_CASE ) and issubclass(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) and inspect.getmodule(_SCREAMING_SNAKE_CASE ) == inspect.getmodule(_config_class ) , ) ] for config_class in config_classes_in_module: lowerCAmelCase__ :Union[str, Any] = check_config_attributes_being_used(_SCREAMING_SNAKE_CASE ) if len(_SCREAMING_SNAKE_CASE ) > 0: lowerCAmelCase__ :int = unused_attributes if len(_SCREAMING_SNAKE_CASE ) > 0: lowerCAmelCase__ :Any = 'The following configuration classes contain unused attributes in the corresponding modeling files:\n' for name, attributes in configs_with_unused_attributes.items(): error += F"{name}: {attributes}\n" raise ValueError(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": check_config_attributes()
254
1
'''simple docstring''' import requests def lowerCAmelCase_ ( snake_case__ , snake_case__ ): '''simple docstring''' A : int = {'''Content-Type''': '''application/json'''} A : List[str] = requests.post(snake_case__ , json={'''text''': message_body} , headers=snake_case__ ) if response.status_code != 200: A : Optional[Any] = ( '''Request to slack returned an error ''' F'{response.status_code}, the response is:\n{response.text}' ) raise ValueError(snake_case__ ) if __name__ == "__main__": # Set the slack url to the one provided by Slack when you create the webhook at # https://my.slack.com/services/new/incoming-webhook/ send_slack_message('<YOUR MESSAGE BODY>', '<SLACK CHANNEL URL>')
3
'''simple docstring''' from typing import List, Optional, Union import numpy as np import torch import torchaudio.compliance.kaldi as ta_kaldi from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging lowercase : Optional[int] = logging.get_logger(__name__) class A ( __snake_case ): __magic_name__ = ['''input_features''', '''attention_mask'''] def __init__( self , SCREAMING_SNAKE_CASE=80 , SCREAMING_SNAKE_CASE=16000 , SCREAMING_SNAKE_CASE=80 , SCREAMING_SNAKE_CASE=0.0 , SCREAMING_SNAKE_CASE=True , SCREAMING_SNAKE_CASE=True , SCREAMING_SNAKE_CASE=True , **SCREAMING_SNAKE_CASE , ) -> int: """simple docstring""" super().__init__(feature_size=SCREAMING_SNAKE_CASE , sampling_rate=SCREAMING_SNAKE_CASE , padding_value=SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) A : Optional[int] = num_mel_bins A : Tuple = do_ceptral_normalize A : Dict = normalize_means A : List[Any] = normalize_vars A : List[str] = True def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , ) -> np.ndarray: """simple docstring""" A : List[Any] = waveform * (2**15) # Kaldi compliance: 16-bit signed integers A : Any = torch.from_numpy(SCREAMING_SNAKE_CASE ).unsqueeze(0 ) A : Any = ta_kaldi.fbank(SCREAMING_SNAKE_CASE , num_mel_bins=self.num_mel_bins , sample_frequency=self.sampling_rate ) return features.numpy() @staticmethod def __lowerCAmelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = True , SCREAMING_SNAKE_CASE = True , SCREAMING_SNAKE_CASE = 0.0 , ) -> np.ndarray: """simple docstring""" if normalize_means: A : Dict = x[:input_length].mean(axis=0 ) A : Optional[Any] = np.subtract(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if normalize_vars: A : str = x[:input_length].std(axis=0 ) A : int = np.divide(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if input_length < x.shape[0]: A : List[str] = padding_value # make sure array is in float32 A : Tuple = x.astype(np.floataa ) return x def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None ) -> List[np.ndarray]: """simple docstring""" A : List[Any] = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [ self.utterance_cmvn(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , self.normalize_means , self.normalize_vars , self.padding_value ) for x, n in zip(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ] def __call__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = False , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = False , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , **SCREAMING_SNAKE_CASE , ) -> BatchFeature: """simple docstring""" if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F'The model corresponding to this feature extractor: {self} was trained using a sampling rate of' F' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with' F' {self.sampling_rate} and not {sampling_rate}.' ) else: logger.warning( '''It is strongly recommended to pass the `sampling_rate` argument to this function. ''' '''Failing to do so can result in silent errors that might be hard to debug.''' ) A : List[Any] = isinstance(SCREAMING_SNAKE_CASE , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F'Only mono-channel audio is supported for input to {self}' ) A : Tuple = is_batched_numpy or ( isinstance(SCREAMING_SNAKE_CASE , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: A : Union[str, Any] = [np.asarray(SCREAMING_SNAKE_CASE , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(SCREAMING_SNAKE_CASE , np.ndarray ): A : Union[str, Any] = np.asarray(SCREAMING_SNAKE_CASE , dtype=np.floataa ) elif isinstance(SCREAMING_SNAKE_CASE , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): A : Optional[int] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: A : Any = [raw_speech] # extract fbank features A : List[str] = [self._extract_fbank_features(SCREAMING_SNAKE_CASE ) for waveform in raw_speech] # convert into correct format for padding A : str = BatchFeature({'''input_features''': features} ) A : Union[str, Any] = self.pad( SCREAMING_SNAKE_CASE , padding=SCREAMING_SNAKE_CASE , max_length=SCREAMING_SNAKE_CASE , truncation=SCREAMING_SNAKE_CASE , pad_to_multiple_of=SCREAMING_SNAKE_CASE , return_attention_mask=SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE , ) # make sure list is in array format A : List[str] = padded_inputs.get('''input_features''' ) if isinstance(input_features[0] , SCREAMING_SNAKE_CASE ): A : str = [np.asarray(SCREAMING_SNAKE_CASE , dtype=np.floataa ) for feature in input_features] A : Union[str, Any] = padded_inputs.get('''attention_mask''' ) if attention_mask is not None: A : Union[str, Any] = [np.asarray(SCREAMING_SNAKE_CASE , dtype=np.intaa ) for array in attention_mask] # Utterance-level cepstral mean and variance normalization if self.do_ceptral_normalize: A : Dict = ( np.array(SCREAMING_SNAKE_CASE , dtype=np.intaa ) if self._get_padding_strategies(SCREAMING_SNAKE_CASE , max_length=SCREAMING_SNAKE_CASE ) is not PaddingStrategy.DO_NOT_PAD else None ) A : List[Any] = self.normalize( padded_inputs['''input_features'''] , attention_mask=SCREAMING_SNAKE_CASE ) if return_tensors is not None: A : int = padded_inputs.convert_to_tensors(SCREAMING_SNAKE_CASE ) return padded_inputs
3
1
import argparse import torch # Step 1. clone https://github.com/microsoft/unilm # Step 2. git checkout to https://github.com/microsoft/unilm/commit/b94ec76c36f02fb2b0bf0dcb0b8554a2185173cd # Step 3. cd unilm # Step 4. ln -s $(realpath wavlm/modules.py) ./ # create simlink # import classes from unilm.wavlm.WavLM import WavLM as WavLMOrig from unilm.wavlm.WavLM import WavLMConfig as WavLMConfigOrig from transformers import WavLMConfig, WavLMModel, logging logging.set_verbosity_info() UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn.grep_linear': 'encoder.layers.*.attention.gru_rel_pos_linear', 'self_attn.relative_attention_bias': 'encoder.layers.*.attention.rel_attn_embed', 'self_attn.grep_a': 'encoder.layers.*.attention.gru_rel_pos_const', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'ctc_proj', 'mask_emb': 'masked_spec_embed', } UpperCAmelCase_ = [ 'ctc_proj', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', ] def lowerCamelCase__ ( A__ : List[Any] , A__ : Optional[int] , A__ : str , A__ : Union[str, Any] , A__ : List[Any] ): '''simple docstring''' for attribute in key.split(""".""" ): __lowerCamelCase = getattr(A__ , A__ ) if weight_type is not None: __lowerCamelCase = getattr(A__ , A__ ).shape else: __lowerCamelCase = hf_pointer.shape assert hf_shape == value.shape, ( f'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' f' {value.shape} for {full_name}' ) if weight_type == "weight": __lowerCamelCase = value elif weight_type == "weight_g": __lowerCamelCase = value elif weight_type == "weight_v": __lowerCamelCase = value elif weight_type == "bias": __lowerCamelCase = value else: __lowerCamelCase = value logger.info(f'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def lowerCamelCase__ ( A__ : Dict , A__ : str ): '''simple docstring''' __lowerCamelCase = [] __lowerCamelCase = fairseq_model.state_dict() __lowerCamelCase = hf_model.feature_extractor for name, value in fairseq_dict.items(): __lowerCamelCase = False if "conv_layers" in name: load_conv_layer( A__ , A__ , A__ , A__ , hf_model.config.feat_extract_norm == """group""" , ) __lowerCamelCase = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split("""w2v_model.""" )[-1] == name.split(""".""" )[0]: __lowerCamelCase = True if "*" in mapped_key: __lowerCamelCase = name.split(A__ )[0].split(""".""" )[-2] __lowerCamelCase = mapped_key.replace("""*""" , A__ ) if "weight_g" in name: __lowerCamelCase = """weight_g""" elif "weight_v" in name: __lowerCamelCase = """weight_v""" elif "bias" in name and "relative_attention_bias" not in name: __lowerCamelCase = """bias""" elif "weight" in name: # TODO: don't match quantizer.weight_proj __lowerCamelCase = """weight""" else: __lowerCamelCase = None set_recursively(A__ , A__ , A__ , A__ , A__ ) continue if not is_used: unused_weights.append(A__ ) logger.warning(f'Unused weights: {unused_weights}' ) def lowerCamelCase__ ( A__ : Any , A__ : int , A__ : str , A__ : Optional[int] , A__ : Optional[Any] ): '''simple docstring''' __lowerCamelCase = full_name.split("""conv_layers.""" )[-1] __lowerCamelCase = name.split(""".""" ) __lowerCamelCase = int(items[0] ) __lowerCamelCase = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) __lowerCamelCase = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) __lowerCamelCase = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) __lowerCamelCase = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f'{full_name} has size {value.shape}, but' f' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) __lowerCamelCase = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(A__ ) @torch.no_grad() def lowerCamelCase__ ( A__ : Dict , A__ : int , A__ : List[str]=None ): '''simple docstring''' __lowerCamelCase = torch.load(A__ ) __lowerCamelCase = WavLMConfigOrig(checkpoint["""cfg"""] ) __lowerCamelCase = WavLMOrig(A__ ) model.load_state_dict(checkpoint["""model"""] ) model.eval() if config_path is not None: __lowerCamelCase = WavLMConfig.from_pretrained(A__ ) else: __lowerCamelCase = WavLMConfig() __lowerCamelCase = WavLMModel(A__ ) recursively_load_weights(A__ , A__ ) hf_wavlm.save_pretrained(A__ ) if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') UpperCAmelCase_ = parser.parse_args() convert_wavlm_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
29
from math import ceil, sqrt def lowerCamelCase__ ( A__ : int = 1000000 ): '''simple docstring''' __lowerCamelCase = 0 for outer_width in range(3 , (limit // 4) + 2 ): if outer_width**2 > limit: __lowerCamelCase = max(ceil(sqrt(outer_width**2 - limit ) ) , 1 ) else: __lowerCamelCase = 1 if (outer_width - hole_width_lower_bound) % 2: hole_width_lower_bound += 1 answer += (outer_width - hole_width_lower_bound - 2) // 2 + 1 return answer if __name__ == "__main__": print(f"""{solution() = }""")
29
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) lowercase__ = { """configuration_blip""": [ """BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BlipConfig""", """BlipTextConfig""", """BlipVisionConfig""", ], """processing_blip""": ["""BlipProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = ["""BlipImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ """BLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """BlipModel""", """BlipPreTrainedModel""", """BlipForConditionalGeneration""", """BlipForQuestionAnswering""", """BlipVisionModel""", """BlipTextModel""", """BlipForImageTextRetrieval""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ """TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFBlipModel""", """TFBlipPreTrainedModel""", """TFBlipForConditionalGeneration""", """TFBlipForQuestionAnswering""", """TFBlipVisionModel""", """TFBlipTextModel""", """TFBlipForImageTextRetrieval""", ] if TYPE_CHECKING: from .configuration_blip import BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipConfig, BlipTextConfig, BlipVisionConfig from .processing_blip import BlipProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_blip import BlipImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip import ( BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, BlipModel, BlipPreTrainedModel, BlipTextModel, BlipVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blip import ( TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFBlipForConditionalGeneration, TFBlipForImageTextRetrieval, TFBlipForQuestionAnswering, TFBlipModel, TFBlipPreTrainedModel, TFBlipTextModel, TFBlipVisionModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
96
'''simple docstring''' import tempfile import unittest import numpy as np from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import BertConfig, is_flax_available from transformers.testing_utils import TOKEN, USER, is_staging_test, require_flax if is_flax_available(): import os from flax.core.frozen_dict import unfreeze from flax.traverse_util import flatten_dict from transformers import FlaxBertModel UpperCamelCase__: Tuple = "0.12" # assumed parallelism: 8 @require_flax @is_staging_test class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" @classmethod def A ( cls : Union[str, Any] ) -> int: UpperCAmelCase : Optional[Any] = TOKEN HfFolder.save_token(__snake_case ) @classmethod def A ( cls : List[str] ) -> Tuple: try: delete_repo(token=cls._token , repo_id='''test-model-flax''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''valid_org/test-model-flax-org''' ) except HTTPError: pass def A ( self : int ) -> Tuple: UpperCAmelCase : List[Any] = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) UpperCAmelCase : Dict = FlaxBertModel(__snake_case ) model.push_to_hub('''test-model-flax''' , use_auth_token=self._token ) UpperCAmelCase : Tuple = FlaxBertModel.from_pretrained(F"""{USER}/test-model-flax""" ) UpperCAmelCase : List[Any] = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : Tuple = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : Union[str, Any] = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) # Reset repo delete_repo(token=self._token , repo_id='''test-model-flax''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(__snake_case , repo_id='''test-model-flax''' , push_to_hub=__snake_case , use_auth_token=self._token ) UpperCAmelCase : str = FlaxBertModel.from_pretrained(F"""{USER}/test-model-flax""" ) UpperCAmelCase : Any = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : str = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : Optional[Any] = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) def A ( self : Optional[Any] ) -> Union[str, Any]: UpperCAmelCase : Dict = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) UpperCAmelCase : Optional[Any] = FlaxBertModel(__snake_case ) model.push_to_hub('''valid_org/test-model-flax-org''' , use_auth_token=self._token ) UpperCAmelCase : Union[str, Any] = FlaxBertModel.from_pretrained('''valid_org/test-model-flax-org''' ) UpperCAmelCase : List[Any] = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : int = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : Any = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) # Reset repo delete_repo(token=self._token , repo_id='''valid_org/test-model-flax-org''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained( __snake_case , repo_id='''valid_org/test-model-flax-org''' , push_to_hub=__snake_case , use_auth_token=self._token ) UpperCAmelCase : str = FlaxBertModel.from_pretrained('''valid_org/test-model-flax-org''' ) UpperCAmelCase : Any = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : Optional[Any] = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : int = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ) -> Union[str, Any]: UpperCAmelCase : str = True UpperCAmelCase : int = flatten_dict(modela.params ) UpperCAmelCase : Dict = flatten_dict(modela.params ) for key in flat_params_a.keys(): if np.sum(np.abs(flat_params_a[key] - flat_params_a[key] ) ) > 1e-4: UpperCAmelCase : Dict = False return models_are_equal @require_flax class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : Tuple ) -> Union[str, Any]: UpperCAmelCase : List[Any] = BertConfig.from_pretrained('''hf-internal-testing/tiny-bert-flax-only''' ) UpperCAmelCase : Dict = FlaxBertModel(__snake_case ) UpperCAmelCase : int = '''bert''' with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(__snake_case , __snake_case ) ) with self.assertRaises(__snake_case ): UpperCAmelCase : Tuple = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : str = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertTrue(check_models_equal(__snake_case , __snake_case ) ) def A ( self : List[str] ) -> Dict: UpperCAmelCase : Dict = BertConfig.from_pretrained('''hf-internal-testing/tiny-bert-flax-only''' ) UpperCAmelCase : Dict = FlaxBertModel(__snake_case ) UpperCAmelCase : Optional[int] = '''bert''' with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(__snake_case , __snake_case ) , max_shard_size='''10KB''' ) with self.assertRaises(__snake_case ): UpperCAmelCase : Any = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : Union[str, Any] = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertTrue(check_models_equal(__snake_case , __snake_case ) ) def A ( self : Optional[int] ) -> str: UpperCAmelCase : Dict = '''bert''' UpperCAmelCase : int = '''hf-internal-testing/tiny-random-bert-subfolder''' with self.assertRaises(__snake_case ): UpperCAmelCase : Optional[Any] = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : Tuple = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertIsNotNone(__snake_case ) def A ( self : Dict ) -> List[Any]: UpperCAmelCase : Optional[int] = '''bert''' UpperCAmelCase : int = '''hf-internal-testing/tiny-random-bert-sharded-subfolder''' with self.assertRaises(__snake_case ): UpperCAmelCase : Dict = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : Union[str, Any] = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertIsNotNone(__snake_case )
23
0
from importlib import import_module from .logging import get_logger lowercase__ :Optional[Any] = get_logger(__name__) class lowercase : def __init__( self ,A__ ,A__=None): lowercase = attrs or [] if module is not None: for key in module.__dict__: if key in attrs or not key.startswith('''__'''): setattr(self ,A__ ,getattr(A__ ,A__)) lowercase = module._original_module if isinstance(A__ ,_PatchedModuleObj) else module class lowercase : lowercase_ : Union[str, Any] =[] def __init__( self ,A__ ,A__ ,A__ ,A__=None): lowercase = obj lowercase = target lowercase = new lowercase = target.split('''.''')[0] lowercase = {} lowercase = attrs or [] def __enter__( self): *lowercase , lowercase = self.target.split('''.''') # Patch modules: # it's used to patch attributes of submodules like "os.path.join"; # in this case we need to patch "os" and "os.path" for i in range(len(A__)): try: lowercase = import_module('''.'''.join(submodules[: i + 1])) except ModuleNotFoundError: continue # We iterate over all the globals in self.obj in case we find "os" or "os.path" for attr in self.obj.__dir__(): lowercase = getattr(self.obj ,A__) # We don't check for the name of the global, but rather if its value *is* "os" or "os.path". # This allows to patch renamed modules like "from os import path as ospath". if obj_attr is submodule or ( (isinstance(A__ ,_PatchedModuleObj) and obj_attr._original_module is submodule) ): lowercase = obj_attr # patch at top level setattr(self.obj ,A__ ,_PatchedModuleObj(A__ ,attrs=self.attrs)) lowercase = getattr(self.obj ,A__) # construct lower levels patches for key in submodules[i + 1 :]: setattr(A__ ,A__ ,_PatchedModuleObj(getattr(A__ ,A__ ,A__) ,attrs=self.attrs)) lowercase = getattr(A__ ,A__) # finally set the target attribute setattr(A__ ,A__ ,self.new) # Patch attribute itself: # it's used for builtins like "open", # and also to patch "os.path.join" we may also need to patch "join" # itself if it was imported as "from os.path import join". if submodules: # if it's an attribute of a submodule like "os.path.join" try: lowercase = getattr(import_module('''.'''.join(A__)) ,A__) except (AttributeError, ModuleNotFoundError): return # We iterate over all the globals in self.obj in case we find "os.path.join" for attr in self.obj.__dir__(): # We don't check for the name of the global, but rather if its value *is* "os.path.join". # This allows to patch renamed attributes like "from os.path import join as pjoin". if getattr(self.obj ,A__) is attr_value: lowercase = getattr(self.obj ,A__) setattr(self.obj ,A__ ,self.new) elif target_attr in globals()["__builtins__"]: # if it'a s builtin like "open" lowercase = globals()['''__builtins__'''][target_attr] setattr(self.obj ,A__ ,self.new) else: raise RuntimeError(f'Tried to patch attribute {target_attr} instead of a submodule.') def __exit__( self ,*A__): for attr in list(self.original): setattr(self.obj ,A__ ,self.original.pop(A__)) def A__ ( self): self.__enter__() self._active_patches.append(self) def A__ ( self): try: self._active_patches.remove(self) except ValueError: # If the patch hasn't been started this will fail return None return self.__exit__()
370
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowercase__ :Any = { "configuration_distilbert": [ "DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "DistilBertConfig", "DistilBertOnnxConfig", ], "tokenization_distilbert": ["DistilBertTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :Tuple = ["DistilBertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :Union[str, Any] = [ "DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "DistilBertForMaskedLM", "DistilBertForMultipleChoice", "DistilBertForQuestionAnswering", "DistilBertForSequenceClassification", "DistilBertForTokenClassification", "DistilBertModel", "DistilBertPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :Optional[int] = [ "TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFDistilBertForMaskedLM", "TFDistilBertForMultipleChoice", "TFDistilBertForQuestionAnswering", "TFDistilBertForSequenceClassification", "TFDistilBertForTokenClassification", "TFDistilBertMainLayer", "TFDistilBertModel", "TFDistilBertPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :Union[str, Any] = [ "FlaxDistilBertForMaskedLM", "FlaxDistilBertForMultipleChoice", "FlaxDistilBertForQuestionAnswering", "FlaxDistilBertForSequenceClassification", "FlaxDistilBertForTokenClassification", "FlaxDistilBertModel", "FlaxDistilBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_distilbert import ( DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DistilBertConfig, DistilBertOnnxConfig, ) from .tokenization_distilbert import DistilBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_distilbert_fast import DistilBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_distilbert import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, DistilBertPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertMainLayer, TFDistilBertModel, TFDistilBertPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, FlaxDistilBertPreTrainedModel, ) else: import sys lowercase__ :List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
97
0
import argparse import torch from ...utils import logging from . import AlbertConfig, AlbertForPreTraining, load_tf_weights_in_albert logging.set_verbosity_info() def __snake_case ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): # Initialise PyTorch model __a = AlbertConfig.from_json_file(_UpperCAmelCase ) print(f'Building PyTorch model from configuration: {config}' ) __a = AlbertForPreTraining(_UpperCAmelCase ) # Load weights from tf checkpoint load_tf_weights_in_albert(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # Save pytorch-model print(f'Save PyTorch model to {pytorch_dump_path}' ) torch.save(model.state_dict() , _UpperCAmelCase ) if __name__ == "__main__": __snake_case :List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--albert_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained ALBERT model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) __snake_case :List[str] = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.albert_config_file, args.pytorch_dump_path)
49
"""simple docstring""" import os import unittest from transformers import FunnelTokenizer, FunnelTokenizerFast from transformers.models.funnel.tokenization_funnel import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = FunnelTokenizer SCREAMING_SNAKE_CASE = FunnelTokenizerFast SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = True def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Optional[int]: """simple docstring""" super().setUp() __lowerCAmelCase : str = [ "<unk>", "<cls>", "<sep>", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] __lowerCAmelCase : int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"]) with open(self.vocab_file , "w" , encoding="utf-8") as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens])) def _SCREAMING_SNAKE_CASE ( self: Optional[Any] , **_SCREAMING_SNAKE_CASE: Union[str, Any]) -> int: """simple docstring""" return FunnelTokenizer.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Any , **_SCREAMING_SNAKE_CASE: Any) -> str: """simple docstring""" return FunnelTokenizerFast.from_pretrained(self.tmpdirname , **_SCREAMING_SNAKE_CASE) def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: str) -> Any: """simple docstring""" __lowerCAmelCase : Union[str, Any] = "UNwant\u00E9d,running" __lowerCAmelCase : str = "unwanted, running" return input_text, output_text def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> List[str]: """simple docstring""" __lowerCAmelCase : Any = self.tokenizer_class(self.vocab_file) __lowerCAmelCase : Any = tokenizer.tokenize("UNwant\u00E9d,running") self.assertListEqual(_SCREAMING_SNAKE_CASE , ["un", "##want", "##ed", ",", "runn", "##ing"]) self.assertListEqual(tokenizer.convert_tokens_to_ids(_SCREAMING_SNAKE_CASE) , [7, 4, 5, 10, 8, 9]) def _SCREAMING_SNAKE_CASE ( self: List[str]) -> List[str]: """simple docstring""" __lowerCAmelCase : Optional[Any] = self.get_tokenizers(do_lower_case=_SCREAMING_SNAKE_CASE) for tokenizer in tokenizers: __lowerCAmelCase : List[str] = tokenizer("UNwant\u00E9d,running") __lowerCAmelCase : Optional[int] = len(inputs["input_ids"]) - 1 self.assertListEqual(inputs["token_type_ids"] , [2] + [0] * sentence_len) __lowerCAmelCase : List[str] = tokenizer("UNwant\u00E9d,running" , "UNwant\u00E9d,running") self.assertListEqual(inputs["token_type_ids"] , [2] + [0] * sentence_len + [1] * sentence_len)
269
0
'''simple docstring''' def _snake_case ( A , A ) -> Union[str, Any]: lowerCAmelCase__ = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): lowerCAmelCase__ = n - k # Calculate C(n,k) for i in range(A ): result *= n - i result //= i + 1 return result def _snake_case ( A ) -> Union[str, Any]: return binomial_coefficient(2 * node_count , A ) // (node_count + 1) def _snake_case ( A ) -> Any: if n < 0: raise ValueError('''factorial() not defined for negative values''' ) lowerCAmelCase__ = 1 for i in range(1 , n + 1 ): result *= i return result def _snake_case ( A ) -> Optional[int]: return catalan_number(A ) * factorial(A ) if __name__ == "__main__": __UpperCAmelCase = int(input('''Enter the number of nodes: ''').strip() or 0) if node_count <= 0: raise ValueError('''We need some nodes to work with.''') print( f"""Given {node_count} nodes, there are {binary_tree_count(node_count)} """ f"""binary trees and {catalan_number(node_count)} binary search trees.""" )
356
'''simple docstring''' import argparse import json import os import re import torch from transformers import BloomConfig, BloomModel from transformers.file_utils import CONFIG_NAME, WEIGHTS_NAME from transformers.utils import logging logging.set_verbosity_info() __UpperCAmelCase = [ '''word_embeddings_layernorm.weight''', '''word_embeddings_layernorm.bias''', '''input_layernorm.weight''', '''input_layernorm.bias''', '''post_attention_layernorm.weight''', '''post_attention_layernorm.bias''', '''self_attention.dense.bias''', '''mlp.dense_4h_to_h.bias''', '''ln_f.weight''', '''ln_f.bias''', ] __UpperCAmelCase = [ '''mlp.dense_4h_to_h.weight''', '''self_attention.dense.weight''', ] def _snake_case ( A , A ) -> Optional[Any]: lowerCAmelCase__ = { '''word_embeddings.weight''': '''word_embeddings.weight''', '''word_embeddings.norm.weight''': '''word_embeddings_layernorm.weight''', '''word_embeddings.norm.bias''': '''word_embeddings_layernorm.bias''', '''weight''': '''ln_f.weight''', '''bias''': '''ln_f.bias''', } if key in layer_rename_map: return layer_rename_map[key] # Handle transformer blocks lowerCAmelCase__ = int(re.match(R'''.*layer_(\d*).*''' , A )[1] ) layer_number -= 3 return F"""h.{layer_number}.""" + key def _snake_case ( A ) -> Optional[int]: if dtype == torch.bool: return 1 / 8 lowerCAmelCase__ = re.search(R'''[^\d](\d+)$''' , str(A ) ) if bit_search is None: raise ValueError(F"""`dtype` is not a valid dtype: {dtype}.""" ) lowerCAmelCase__ = int(bit_search.groups()[0] ) return bit_size // 8 def _snake_case ( A , A , A , A , A ) -> Dict: # Construct model if bloom_config_file == "": lowerCAmelCase__ = BloomConfig() else: lowerCAmelCase__ = BloomConfig.from_json_file(A ) if shard_model: lowerCAmelCase__ = os.listdir(A ) lowerCAmelCase__ = sorted(filter(lambda A : s.startswith('''layer''' ) and "model_00" in s , A ) ) lowerCAmelCase__ = {'''weight_map''': {}, '''metadata''': {}} lowerCAmelCase__ = 0 lowerCAmelCase__ = None lowerCAmelCase__ = BloomConfig() for j, file in enumerate(A ): print('''Processing file: {}'''.format(A ) ) lowerCAmelCase__ = None for i in range(A ): # load all TP files lowerCAmelCase__ = file.replace('''model_00''' , F"""model_0{i}""" ) lowerCAmelCase__ = torch.load(os.path.join(A , A ) , map_location='''cpu''' ) # Rename keys in the transformers names lowerCAmelCase__ = list(temp.keys() ) for key in keys: lowerCAmelCase__ = temp.pop(A ) if tensors is None: lowerCAmelCase__ = temp else: for key in tensors.keys(): if any(key.endswith(A ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): # We average (sum and then divide) some weights accross TP ranks (see https://github.com/bigscience-workshop/Megatron-DeepSpeed/blob/olruwase/sync_layer_norms/megatron/training.py#L425) tensors[key] += temp[key] else: # Some weights are RowParallelLinear in Megatron-Deepspeed, others are ColumnParallel lowerCAmelCase__ = 1 if any(text in key for text in WEIGHTS_WITH_ROW_PARALLELISM_CONTAIN ) else 0 # We concatenate these weights accross TP ranks lowerCAmelCase__ = torch.cat([tensors[key], temp[key]] , dim=A ) # Divide by the number of TP the weights we want to average for key in tensors.keys(): if any(key.endswith(A ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): lowerCAmelCase__ = tensors[key] / pretraining_tp torch.save( A , os.path.join( A , '''pytorch_model_{}-of-{}.bin'''.format(str(j + 1 ).zfill(5 ) , str(len(A ) ).zfill(5 ) ) , ) , ) for key in tensors.keys(): lowerCAmelCase__ = tensors[key] total_size += value.numel() * get_dtype_size(value.dtype ) if key not in index_dict["weight_map"]: lowerCAmelCase__ = '''pytorch_model_{}-of-{}.bin'''.format( str(j + 1 ).zfill(5 ) , str(len(A ) ).zfill(5 ) ) lowerCAmelCase__ = BloomConfig() lowerCAmelCase__ = pytorch_dump_folder_path + '''/''' + CONFIG_NAME lowerCAmelCase__ = total_size with open(A , '''w''' , encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) with open(os.path.join(A , WEIGHTS_NAME + '''.index.json''' ) , '''w''' , encoding='''utf-8''' ) as f: lowerCAmelCase__ = json.dumps(A , indent=2 , sort_keys=A ) + '''\n''' f.write(A ) else: lowerCAmelCase__ = BloomModel(A ) lowerCAmelCase__ = os.listdir(A ) lowerCAmelCase__ = sorted(filter(lambda A : s.startswith('''layer''' ) and "model_00" in s , A ) ) lowerCAmelCase__ = None for i, file in enumerate(A ): lowerCAmelCase__ = None for i in range(A ): # load all TP files lowerCAmelCase__ = file.replace('''model_00''' , F"""model_0{i}""" ) lowerCAmelCase__ = torch.load(os.path.join(A , A ) , map_location='''cpu''' ) # Rename keys in the transformers names lowerCAmelCase__ = list(temp.keys() ) for key in keys: lowerCAmelCase__ = temp.pop(A ) if tensors is None: lowerCAmelCase__ = temp else: for key in tensors.keys(): # We average (sum and then divide) some weights accross TP ranks (see https://github.com/bigscience-workshop/Megatron-DeepSpeed/blob/olruwase/sync_layer_norms/megatron/training.py#L425) if any(key.endswith(A ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): tensors[key] += temp[key] else: # Some weights are RowParallelLinear in Megatron-Deepspeed, others are ColumnParallel lowerCAmelCase__ = 1 if any(text in key for text in WEIGHTS_WITH_ROW_PARALLELISM_CONTAIN ) else 0 # We concatenate these weights accross TP ranks lowerCAmelCase__ = torch.cat([tensors[key], temp[key]] , dim=A ) # Divide by the number of TP the weights we want to average for key in tensors.keys(): if any(key.endswith(A ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): lowerCAmelCase__ = tensors[key] / pretraining_tp lowerCAmelCase__ = model.load_state_dict(A , strict=A ) assert not other_keys.unexpected_keys, F"""The keys {other_keys.unexpected_keys} are unexpected""" if missing_keys is None: lowerCAmelCase__ = set(other_keys.missing_keys ) else: lowerCAmelCase__ = missing_keys.intersection(set(other_keys.missing_keys ) ) assert not missing_keys, F"""The keys {missing_keys} are missing""" # Save pytorch-model os.makedirs(A , exist_ok=A ) lowerCAmelCase__ = pytorch_dump_folder_path + '''/''' + WEIGHTS_NAME lowerCAmelCase__ = pytorch_dump_folder_path + '''/''' + CONFIG_NAME print(F"""Save PyTorch model to {pytorch_weights_dump_path} with dtype {config.torch_dtype}""" ) if config.torch_dtype is not None: lowerCAmelCase__ = model.to(config.torch_dtype ) torch.save(model.state_dict() , A ) print(F"""Save configuration file to {pytorch_config_dump_path}""" ) with open(A , '''w''' , encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--bloom_checkpoint_path''', default=None, type=str, required=True, help='''Path to the Megatron-LM checkpoint path.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--bloom_config_file''', default='''''', type=str, help=( '''An optional config json file corresponding to the pre-trained model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--shard_model''', action='''store_true''', help='''An optional setting to shard the output model \nThis enables sharding the converted checkpoint''', ) parser.add_argument( '''--pretraining_tp''', default=4, type=int, help='''Pretraining TP rank that has been used when training the model in Megatron-LM \n''', ) __UpperCAmelCase = parser.parse_args() convert_bloom_checkpoint_to_pytorch( args.bloom_checkpoint_path, args.bloom_config_file, args.pytorch_dump_folder_path, args.shard_model, args.pretraining_tp, )
228
0
'''simple docstring''' import argparse import json import subprocess def _UpperCamelCase ( __A , __A ) -> List[Any]: '''simple docstring''' UpperCamelCase__ = [] UpperCamelCase__ = ( F'''curl -H "Accept: application/vnd.github+json" -H "Authorization: Bearer {token}"''' " https://api.github.com/repos/huggingface/transformers/actions/runners" ) UpperCamelCase__ = subprocess.run(__A , shell=__A , stdout=subprocess.PIPE ) UpperCamelCase__ = output.stdout.decode("utf-8" ) UpperCamelCase__ = json.loads(__A ) UpperCamelCase__ = status["runners"] for runner in runners: if runner["name"] in target_runners: if runner["status"] == "offline": offline_runners.append(__A ) # save the result so we can report them on Slack with open("offline_runners.txt" , "w" ) as fp: fp.write(json.dumps(__A ) ) if len(__A ) > 0: UpperCamelCase__ = "\n".join([x["name"] for x in offline_runners] ) raise ValueError(F'''The following runners are offline:\n{failed}''' ) if __name__ == "__main__": def _UpperCamelCase ( __A ) -> Union[str, Any]: '''simple docstring''' return values.split("," ) a__ : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( '--target_runners', default=None, type=list_str, required=True, help='Comma-separated list of runners to check status.', ) parser.add_argument( '--token', default=None, type=str, required=True, help='A token that has actions:read permission.' ) a__ : Optional[Any] = parser.parse_args() get_runner_status(args.target_runners, args.token)
80
'''simple docstring''' from __future__ import annotations import requests def UpperCamelCase_( snake_case : str ): '''simple docstring''' snake_case_ = f'https://hacker-news.firebaseio.com/v0/item/{story_id}.json?print=pretty' return requests.get(snake_case ).json() def UpperCamelCase_( snake_case : int = 1_0 ): '''simple docstring''' snake_case_ = "https://hacker-news.firebaseio.com/v0/topstories.json?print=pretty" snake_case_ = requests.get(snake_case ).json()[:max_stories] return [get_hackernews_story(snake_case ) for story_id in story_ids] def UpperCamelCase_( snake_case : int = 1_0 ): '''simple docstring''' snake_case_ = hackernews_top_stories(snake_case ) return "\n".join("* [{title}]({url})".format(**snake_case ) for story in stories ) if __name__ == "__main__": print(hackernews_top_stories_as_markdown())
85
0
from typing import Dict from .base import GenericTensor, Pipeline class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def lowercase ( self : Union[str, Any] , lowerCAmelCase_ : Any=None , lowerCAmelCase_ : int=None , lowerCAmelCase_ : Tuple=None , **lowerCAmelCase_ : Optional[Any] ) -> Optional[Any]: if tokenize_kwargs is None: __lowerCAmelCase = {} if truncation is not None: if "truncation" in tokenize_kwargs: raise ValueError( 'truncation parameter defined twice (given as keyword argument as well as in tokenize_kwargs)' ) __lowerCAmelCase = truncation __lowerCAmelCase = tokenize_kwargs __lowerCAmelCase = {} if return_tensors is not None: __lowerCAmelCase = return_tensors return preprocess_params, {}, postprocess_params def lowercase ( self : List[Any] , lowerCAmelCase_ : Optional[Any] , **lowerCAmelCase_ : str ) -> Dict[str, GenericTensor]: __lowerCAmelCase = self.framework __lowerCAmelCase = self.tokenizer(lowerCAmelCase_ , return_tensors=lowerCAmelCase_ , **lowerCAmelCase_ ) return model_inputs def lowercase ( self : int , lowerCAmelCase_ : Optional[int] ) -> int: __lowerCAmelCase = self.model(**lowerCAmelCase_ ) return model_outputs def lowercase ( self : Any , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : int=False ) -> Tuple: # [0] is the first available tensor, logits or last_hidden_state. if return_tensors: return model_outputs[0] if self.framework == "pt": return model_outputs[0].tolist() elif self.framework == "tf": return model_outputs[0].numpy().tolist() def __call__( self : Dict , *lowerCAmelCase_ : Dict , **lowerCAmelCase_ : List[str] ) -> str: return super().__call__(*lowerCAmelCase_ , **lowerCAmelCase_ )
359
from functools import lru_cache @lru_cache def a_ ( lowerCAmelCase_ : int ): if num < 0: raise ValueError('Number should not be negative.' ) return 1 if num in (0, 1) else num * factorial(num - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
207
0
import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) class a ( __lowerCAmelCase ): """simple docstring""" lowerCamelCase :Optional[int] = '''AutoTokenizer''' lowerCamelCase :Tuple = ['''tokenizer'''] lowerCamelCase :Union[str, Any] = { '''semantic_prompt''': 1, '''coarse_prompt''': 2, '''fine_prompt''': 2, } def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=None ) -> List[str]: super().__init__(lowerCAmelCase_ ) _A = speaker_embeddings @classmethod def UpperCAmelCase ( cls , lowerCAmelCase_ , lowerCAmelCase_="speaker_embeddings_path.json" , **lowerCAmelCase_ ) -> Any: if speaker_embeddings_dict_path is not None: _A = get_file_from_repo( lowerCAmelCase_ , lowerCAmelCase_ , subfolder=kwargs.pop("""subfolder""" , lowerCAmelCase_ ) , cache_dir=kwargs.pop("""cache_dir""" , lowerCAmelCase_ ) , force_download=kwargs.pop("""force_download""" , lowerCAmelCase_ ) , proxies=kwargs.pop("""proxies""" , lowerCAmelCase_ ) , resume_download=kwargs.pop("""resume_download""" , lowerCAmelCase_ ) , local_files_only=kwargs.pop("""local_files_only""" , lowerCAmelCase_ ) , use_auth_token=kwargs.pop("""use_auth_token""" , lowerCAmelCase_ ) , revision=kwargs.pop("""revision""" , lowerCAmelCase_ ) , ) if speaker_embeddings_path is None: logger.warning( F'''`{os.path.join(lowerCAmelCase_ , lowerCAmelCase_ )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.''' ) _A = None else: with open(lowerCAmelCase_ ) as speaker_embeddings_json: _A = json.load(lowerCAmelCase_ ) else: _A = None _A = AutoTokenizer.from_pretrained(lowerCAmelCase_ , **lowerCAmelCase_ ) return cls(tokenizer=lowerCAmelCase_ , speaker_embeddings=lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_="speaker_embeddings_path.json" , lowerCAmelCase_="speaker_embeddings" , lowerCAmelCase_ = False , **lowerCAmelCase_ , ) -> str: if self.speaker_embeddings is not None: os.makedirs(os.path.join(lowerCAmelCase_ , lowerCAmelCase_ , """v2""" ) , exist_ok=lowerCAmelCase_ ) _A = {} _A = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": _A = self._load_voice_preset(lowerCAmelCase_ ) _A = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict["""repo_or_path"""] , lowerCAmelCase_ , F'''{prompt_key}_{key}''' ) , voice_preset[key] , allow_pickle=lowerCAmelCase_ , ) _A = os.path.join(lowerCAmelCase_ , F'''{prompt_key}_{key}.npy''' ) _A = tmp_dict with open(os.path.join(lowerCAmelCase_ , lowerCAmelCase_ ) , """w""" ) as fp: json.dump(lowerCAmelCase_ , lowerCAmelCase_ ) super().save_pretrained(lowerCAmelCase_ , lowerCAmelCase_ , **lowerCAmelCase_ ) def UpperCAmelCase ( self , lowerCAmelCase_ = None , **lowerCAmelCase_ ) -> Dict: _A = self.speaker_embeddings[voice_preset] _A = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( F'''Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].''' ) _A = get_file_from_repo( self.speaker_embeddings.get("""repo_or_path""" , """/""" ) , voice_preset_paths[key] , subfolder=kwargs.pop("""subfolder""" , lowerCAmelCase_ ) , cache_dir=kwargs.pop("""cache_dir""" , lowerCAmelCase_ ) , force_download=kwargs.pop("""force_download""" , lowerCAmelCase_ ) , proxies=kwargs.pop("""proxies""" , lowerCAmelCase_ ) , resume_download=kwargs.pop("""resume_download""" , lowerCAmelCase_ ) , local_files_only=kwargs.pop("""local_files_only""" , lowerCAmelCase_ ) , use_auth_token=kwargs.pop("""use_auth_token""" , lowerCAmelCase_ ) , revision=kwargs.pop("""revision""" , lowerCAmelCase_ ) , ) if path is None: raise ValueError( F'''`{os.path.join(self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.''' ) _A = np.load(lowerCAmelCase_ ) return voice_preset_dict def UpperCAmelCase ( self , lowerCAmelCase_ = None ) -> Optional[int]: for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(F'''Voice preset unrecognized, missing {key} as a key.''' ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(F'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(F'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) def __call__( self , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_="pt" , lowerCAmelCase_=2_56 , lowerCAmelCase_=False , lowerCAmelCase_=True , lowerCAmelCase_=False , **lowerCAmelCase_ , ) -> Union[str, Any]: if voice_preset is not None and not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): if ( isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): _A = self._load_voice_preset(lowerCAmelCase_ ) else: if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and not voice_preset.endswith(""".npz""" ): _A = voice_preset + """.npz""" _A = np.load(lowerCAmelCase_ ) if voice_preset is not None: self._validate_voice_preset_dict(lowerCAmelCase_ , **lowerCAmelCase_ ) _A = BatchFeature(data=lowerCAmelCase_ , tensor_type=lowerCAmelCase_ ) _A = self.tokenizer( lowerCAmelCase_ , return_tensors=lowerCAmelCase_ , padding="""max_length""" , max_length=lowerCAmelCase_ , return_attention_mask=lowerCAmelCase_ , return_token_type_ids=lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ , **lowerCAmelCase_ , ) if voice_preset is not None: _A = voice_preset return encoded_text
180
import math import sys def snake_case ( snake_case__ :int) -> int: if number != int(snake_case__): raise ValueError("""the value of input must be a natural number""") if number < 0: raise ValueError("""the value of input must not be a negative number""") if number == 0: return 1 _A = [-1] * (number + 1) _A = 0 for i in range(1 , number + 1): _A = sys.maxsize _A = int(math.sqrt(snake_case__)) for j in range(1 , root + 1): _A = 1 + answers[i - (j**2)] _A = min(snake_case__ , snake_case__) _A = answer return answers[number] if __name__ == "__main__": import doctest doctest.testmod()
180
1
"""simple docstring""" from torch import nn class _SCREAMING_SNAKE_CASE( nn.Module ): def __init__( self ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) -> List[str]: """simple docstring""" super().__init__() __SCREAMING_SNAKE_CASE :Tuple = class_size __SCREAMING_SNAKE_CASE :str = embed_size # self.mlp1 = nn.Linear(embed_size, embed_size) # self.mlp2 = (nn.Linear(embed_size, class_size)) __SCREAMING_SNAKE_CASE :Optional[Any] = nn.Linear(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__ ) -> Optional[int]: """simple docstring""" __SCREAMING_SNAKE_CASE :Optional[Any] = self.mlp(SCREAMING_SNAKE_CASE__ ) return logits
355
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = { "google/vivit-b-16x2-kinetics400": ( "https://huggingface.co/google/vivit-b-16x2-kinetics400/resolve/main/config.json" ), # See all Vivit models at https://huggingface.co/models?filter=vivit } class _SCREAMING_SNAKE_CASE( A ): SCREAMING_SNAKE_CASE_ : Optional[Any] = '''vivit''' def __init__( self ,SCREAMING_SNAKE_CASE__=2_24 ,SCREAMING_SNAKE_CASE__=32 ,SCREAMING_SNAKE_CASE__=[2, 16, 16] ,SCREAMING_SNAKE_CASE__=3 ,SCREAMING_SNAKE_CASE__=7_68 ,SCREAMING_SNAKE_CASE__=12 ,SCREAMING_SNAKE_CASE__=12 ,SCREAMING_SNAKE_CASE__=30_72 ,SCREAMING_SNAKE_CASE__="gelu_fast" ,SCREAMING_SNAKE_CASE__=0.0 ,SCREAMING_SNAKE_CASE__=0.0 ,SCREAMING_SNAKE_CASE__=0.0_2 ,SCREAMING_SNAKE_CASE__=1E-06 ,SCREAMING_SNAKE_CASE__=True ,**SCREAMING_SNAKE_CASE__ ,) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE :Optional[Any] = hidden_size __SCREAMING_SNAKE_CASE :List[str] = num_hidden_layers __SCREAMING_SNAKE_CASE :List[str] = num_attention_heads __SCREAMING_SNAKE_CASE :Any = intermediate_size __SCREAMING_SNAKE_CASE :Optional[Any] = hidden_act __SCREAMING_SNAKE_CASE :int = hidden_dropout_prob __SCREAMING_SNAKE_CASE :Union[str, Any] = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE :Dict = initializer_range __SCREAMING_SNAKE_CASE :Dict = layer_norm_eps __SCREAMING_SNAKE_CASE :Any = image_size __SCREAMING_SNAKE_CASE :Any = num_frames __SCREAMING_SNAKE_CASE :Any = tubelet_size __SCREAMING_SNAKE_CASE :Tuple = num_channels __SCREAMING_SNAKE_CASE :List[str] = qkv_bias super().__init__(**SCREAMING_SNAKE_CASE__ )
239
0
"""simple docstring""" snake_case__ : Optional[Any] = {str(digit): digit**5 for digit in range(10)} def _snake_case ( _snake_case : int ): return sum(DIGITS_FIFTH_POWER[digit] for digit in str(_snake_case ) ) def _snake_case ( ): return sum( number for number in range(1000 , 1000000 ) if number == digits_fifth_powers_sum(_snake_case ) ) if __name__ == "__main__": print(solution())
60
import inspect import logging import os import random import shutil import tempfile import unittest import pytest import torch from torch import nn from torch.utils.data import DataLoader, TensorDataset from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_cuda from accelerate.utils import ProjectConfiguration, set_seed lowercase__ : Optional[Any] = logging.getLogger(__name__) def A_ ( snake_case : Any=2 , snake_case : Union[str, Any]=3 , snake_case : Union[str, Any]=16 , snake_case : int = 10 , snake_case : int = 2 ) -> int: '''simple docstring''' def get_dataset(snake_case : Optional[int] ): __UpperCamelCase = torch.randn(batch_size * n_batches , 1 ) return TensorDataset(snake_case , a * x + b + 0.1 * torch.randn(batch_size * n_batches , 1 ) ) __UpperCamelCase = get_dataset(snake_case ) __UpperCamelCase = get_dataset(snake_case ) __UpperCamelCase = DataLoader(snake_case , shuffle=snake_case , batch_size=snake_case , num_workers=4 ) __UpperCamelCase = DataLoader(snake_case , shuffle=snake_case , batch_size=snake_case , num_workers=4 ) return (train_dataloader, valid_dataloader) def A_ ( snake_case : List[str] , snake_case : int , snake_case : List[str] , snake_case : Optional[int] , snake_case : int , snake_case : str=None ) -> Any: '''simple docstring''' __UpperCamelCase = [] for epoch in range(snake_case ): # Train quickly model.train() for batch in dataloader: __UpperCamelCase , __UpperCamelCase = batch __UpperCamelCase = model(snake_case ) __UpperCamelCase = torch.nn.functional.mse_loss(snake_case , snake_case ) accelerator.backward(snake_case ) optimizer.step() optimizer.zero_grad() rands.append(random.random() ) # Introduce some randomness if scheduler is not None: scheduler.step() return rands class SCREAMING_SNAKE_CASE__ ( nn.Module ): """simple docstring""" def __init__( self )-> Tuple: '''simple docstring''' super().__init__() __UpperCamelCase = nn.Parameter(torch.randn(1 ) ) __UpperCamelCase = nn.Parameter(torch.randn(1 ) ) def A__ ( self , SCREAMING_SNAKE_CASE_ )-> Dict: '''simple docstring''' return x * self.a + self.b class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): """simple docstring""" def A__ ( self )-> Tuple: '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) __UpperCamelCase = DummyModel() __UpperCamelCase = torch.optim.Adam(params=model.parameters() , lr=1E-3 ) __UpperCamelCase , __UpperCamelCase = dummy_dataloaders() __UpperCamelCase = ProjectConfiguration(total_limit=1 , project_dir=SCREAMING_SNAKE_CASE_ , automatic_checkpoint_naming=SCREAMING_SNAKE_CASE_ ) # Train baseline __UpperCamelCase = Accelerator(project_config=SCREAMING_SNAKE_CASE_ ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = accelerator.prepare( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Save initial accelerator.save_state() # Save second state accelerator.save_state() self.assertEqual(len(os.listdir(accelerator.project_dir ) ) , 1 ) def A__ ( self )-> Optional[int]: '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) __UpperCamelCase = DummyModel() __UpperCamelCase = torch.optim.Adam(params=model.parameters() , lr=1E-3 ) __UpperCamelCase , __UpperCamelCase = dummy_dataloaders() # Train baseline __UpperCamelCase = Accelerator() __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = accelerator.prepare( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Save initial __UpperCamelCase = os.path.join(SCREAMING_SNAKE_CASE_ , '''initial''' ) accelerator.save_state(SCREAMING_SNAKE_CASE_ ) ((__UpperCamelCase) , (__UpperCamelCase)) = model.a.item(), model.b.item() __UpperCamelCase = optimizer.state_dict() __UpperCamelCase = train(3 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ((__UpperCamelCase) , (__UpperCamelCase)) = model.a.item(), model.b.item() __UpperCamelCase = optimizer.state_dict() # Train partially set_seed(42 ) __UpperCamelCase = DummyModel() __UpperCamelCase = torch.optim.Adam(params=model.parameters() , lr=1E-3 ) __UpperCamelCase , __UpperCamelCase = dummy_dataloaders() __UpperCamelCase = Accelerator() __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = accelerator.prepare( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) accelerator.load_state(SCREAMING_SNAKE_CASE_ ) ((__UpperCamelCase) , (__UpperCamelCase)) = model.a.item(), model.b.item() __UpperCamelCase = optimizer.state_dict() self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) __UpperCamelCase = train(2 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Save everything __UpperCamelCase = os.path.join(SCREAMING_SNAKE_CASE_ , '''checkpoint''' ) accelerator.save_state(SCREAMING_SNAKE_CASE_ ) # Load everything back in and make sure all states work accelerator.load_state(SCREAMING_SNAKE_CASE_ ) test_rands += train(1 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ((__UpperCamelCase) , (__UpperCamelCase)) = model.a.item(), model.b.item() __UpperCamelCase = optimizer.state_dict() self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def A__ ( self )-> List[str]: '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) __UpperCamelCase = DummyModel() __UpperCamelCase = torch.optim.Adam(params=model.parameters() , lr=1E-3 ) __UpperCamelCase , __UpperCamelCase = dummy_dataloaders() __UpperCamelCase = ProjectConfiguration(automatic_checkpoint_naming=SCREAMING_SNAKE_CASE_ ) # Train baseline __UpperCamelCase = Accelerator(project_dir=SCREAMING_SNAKE_CASE_ , project_config=SCREAMING_SNAKE_CASE_ ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = accelerator.prepare( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Save initial accelerator.save_state() ((__UpperCamelCase) , (__UpperCamelCase)) = model.a.item(), model.b.item() __UpperCamelCase = optimizer.state_dict() __UpperCamelCase = train(3 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ((__UpperCamelCase) , (__UpperCamelCase)) = model.a.item(), model.b.item() __UpperCamelCase = optimizer.state_dict() # Train partially set_seed(42 ) __UpperCamelCase = DummyModel() __UpperCamelCase = torch.optim.Adam(params=model.parameters() , lr=1E-3 ) __UpperCamelCase , __UpperCamelCase = dummy_dataloaders() __UpperCamelCase = ProjectConfiguration(iteration=1 , automatic_checkpoint_naming=SCREAMING_SNAKE_CASE_ ) __UpperCamelCase = Accelerator(project_dir=SCREAMING_SNAKE_CASE_ , project_config=SCREAMING_SNAKE_CASE_ ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = accelerator.prepare( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) accelerator.load_state(os.path.join(SCREAMING_SNAKE_CASE_ , '''checkpoints''' , '''checkpoint_0''' ) ) ((__UpperCamelCase) , (__UpperCamelCase)) = model.a.item(), model.b.item() __UpperCamelCase = optimizer.state_dict() self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) __UpperCamelCase = train(2 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Save everything accelerator.save_state() # Load everything back in and make sure all states work accelerator.load_state(os.path.join(SCREAMING_SNAKE_CASE_ , '''checkpoints''' , '''checkpoint_1''' ) ) test_rands += train(1 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ((__UpperCamelCase) , (__UpperCamelCase)) = model.a.item(), model.b.item() __UpperCamelCase = optimizer.state_dict() self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def A__ ( self )-> Tuple: '''simple docstring''' __UpperCamelCase = torch.tensor([1, 2, 3] ) __UpperCamelCase = torch.tensor([2, 3, 4] ) __UpperCamelCase = DummyModel() __UpperCamelCase = torch.optim.Adam(net.parameters() ) __UpperCamelCase = Accelerator() with self.assertRaises(SCREAMING_SNAKE_CASE_ ) as ve: accelerator.register_for_checkpointing(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) __UpperCamelCase = str(ve.exception ) self.assertTrue('''Item at index 0''' in message ) self.assertTrue('''Item at index 1''' in message ) self.assertFalse('''Item at index 2''' in message ) self.assertFalse('''Item at index 3''' in message ) def A__ ( self )-> Union[str, Any]: '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) __UpperCamelCase = DummyModel() __UpperCamelCase = torch.optim.Adam(params=model.parameters() , lr=1E-3 ) __UpperCamelCase = torch.optim.lr_scheduler.StepLR(SCREAMING_SNAKE_CASE_ , step_size=1 , gamma=0.9_9 ) __UpperCamelCase , __UpperCamelCase = dummy_dataloaders() __UpperCamelCase = ProjectConfiguration(automatic_checkpoint_naming=SCREAMING_SNAKE_CASE_ ) # Train baseline __UpperCamelCase = Accelerator(project_dir=SCREAMING_SNAKE_CASE_ , project_config=SCREAMING_SNAKE_CASE_ ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = accelerator.prepare( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # Save initial accelerator.save_state() __UpperCamelCase = scheduler.state_dict() train(3 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) self.assertNotEqual(SCREAMING_SNAKE_CASE_ , scheduler.state_dict() ) # Load everything back in and make sure all states work accelerator.load_state(os.path.join(SCREAMING_SNAKE_CASE_ , '''checkpoints''' , '''checkpoint_0''' ) ) self.assertEqual(SCREAMING_SNAKE_CASE_ , scheduler.state_dict() ) def A__ ( self )-> List[str]: '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) __UpperCamelCase = DummyModel() __UpperCamelCase = ProjectConfiguration(automatic_checkpoint_naming=SCREAMING_SNAKE_CASE_ , total_limit=2 ) # Train baseline __UpperCamelCase = Accelerator(project_dir=SCREAMING_SNAKE_CASE_ , project_config=SCREAMING_SNAKE_CASE_ ) __UpperCamelCase = accelerator.prepare(SCREAMING_SNAKE_CASE_ ) # Save 3 states: for _ in range(11 ): accelerator.save_state() self.assertTrue(not os.path.exists(os.path.join(SCREAMING_SNAKE_CASE_ , '''checkpoints''' , '''checkpoint_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(SCREAMING_SNAKE_CASE_ , '''checkpoints''' , '''checkpoint_9''' ) ) ) self.assertTrue(os.path.exists(os.path.join(SCREAMING_SNAKE_CASE_ , '''checkpoints''' , '''checkpoint_10''' ) ) ) @require_cuda def A__ ( self )-> Optional[int]: '''simple docstring''' __UpperCamelCase = ['''torchrun''', F"--nproc_per_node={torch.cuda.device_count()}", inspect.getfile(self.__class__ )] execute_subprocess_async(SCREAMING_SNAKE_CASE_ , env=os.environ.copy() ) if __name__ == "__main__": lowercase__ : Optional[int] = "/tmp/accelerate/state_checkpointing" lowercase__ : List[Any] = DummyModel() lowercase__ : Tuple = torch.optim.Adam(params=model.parameters(), lr=1e-3) lowercase__ : int = torch.optim.lr_scheduler.StepLR(optimizer, step_size=1, gamma=0.99) lowercase__ , lowercase__ : str = dummy_dataloaders() lowercase__ : Union[str, Any] = ProjectConfiguration(automatic_checkpoint_naming=True) # Train baseline lowercase__ : List[str] = Accelerator(project_dir=savedir, project_config=project_config, mixed_precision="no") if accelerator.process_index == 0: if os.path.exists(savedir): shutil.rmtree(savedir) os.makedirs(savedir) lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ : Dict = accelerator.prepare( model, optimizer, train_dataloader, valid_dataloader, scheduler ) lowercase__ , lowercase__ : str = accelerator.prepare(model, optimizer) train(3, model, train_dataloader, optimizer, accelerator, scheduler) # Check that the intial optimizer is loaded on the GPU for group in optimizer.param_groups: lowercase__ : int = group["params"][0].device break assert param_device.type == accelerator.device.type lowercase__ : Union[str, Any] = model.cpu() accelerator.wait_for_everyone() accelerator.save_state() accelerator.wait_for_everyone() # Check CPU state accelerator.load_state(os.path.join(savedir, "checkpoints", "checkpoint_0"), map_location="cpu") for group in optimizer.param_groups: lowercase__ : Any = group["params"][0].device break assert ( param_device.type == torch.device("cpu").type ), F"Loaded optimizer states did not match, expected to be loaded on the CPU but got {param_device}" # Check device state model.to(accelerator.device) accelerator.load_state(os.path.join(savedir, "checkpoints", "checkpoint_0"), map_location="on_device") for group in optimizer.param_groups: lowercase__ : List[Any] = group["params"][0].device break assert ( param_device.type == accelerator.device.type ), F"Loaded optimizer states did not match, expected to be loaded on {accelerator.device} but got {param_device}" # Check error with pytest.raises(TypeError, match="Unsupported optimizer map location passed"): accelerator.load_state(os.path.join(savedir, "checkpoints", "checkpoint_0"), map_location="invalid") accelerator.wait_for_everyone() if accelerator.process_index == 0: shutil.rmtree(savedir) accelerator.wait_for_everyone()
328
0
'''simple docstring''' import unittest from datasets import load_dataset from transformers.pipelines import pipeline from transformers.testing_utils import is_pipeline_test, nested_simplify, require_torch, slow @is_pipeline_test @require_torch class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @require_torch def _lowerCAmelCase( self ) -> int: lowercase__ : str = pipeline( task='''zero-shot-audio-classification''' , model='''hf-internal-testing/tiny-clap-htsat-unfused''' ) lowercase__ : Optional[Any] = load_dataset('''ashraq/esc50''' ) lowercase__ : Tuple = dataset['''train''']['''audio'''][-1]['''array'''] lowercase__ : Optional[Any] = audio_classifier(__lowerCAmelCase , candidate_labels=['''Sound of a dog''', '''Sound of vaccum cleaner'''] ) self.assertEqual( nested_simplify(__lowerCAmelCase ) , [{'''score''': 0.5_0_1, '''label''': '''Sound of a dog'''}, {'''score''': 0.4_9_9, '''label''': '''Sound of vaccum cleaner'''}] , ) @unittest.skip('''No models are available in TF''' ) def _lowerCAmelCase( self ) -> str: pass @slow @require_torch def _lowerCAmelCase( self ) -> Tuple: lowercase__ : List[str] = pipeline( task='''zero-shot-audio-classification''' , model='''laion/clap-htsat-unfused''' , ) # This is an audio of a dog lowercase__ : int = load_dataset('''ashraq/esc50''' ) lowercase__ : str = dataset['''train''']['''audio'''][-1]['''array'''] lowercase__ : Any = audio_classifier(__lowerCAmelCase , candidate_labels=['''Sound of a dog''', '''Sound of vaccum cleaner'''] ) self.assertEqual( nested_simplify(__lowerCAmelCase ) , [ {'''score''': 0.9_9_9, '''label''': '''Sound of a dog'''}, {'''score''': 0.0_0_1, '''label''': '''Sound of vaccum cleaner'''}, ] , ) lowercase__ : Dict = audio_classifier([audio] * 5 , candidate_labels=['''Sound of a dog''', '''Sound of vaccum cleaner'''] ) self.assertEqual( nested_simplify(__lowerCAmelCase ) , [ [ {'''score''': 0.9_9_9, '''label''': '''Sound of a dog'''}, {'''score''': 0.0_0_1, '''label''': '''Sound of vaccum cleaner'''}, ], ] * 5 , ) lowercase__ : Any = audio_classifier( [audio] * 5 , candidate_labels=['''Sound of a dog''', '''Sound of vaccum cleaner'''] , batch_size=5 ) self.assertEqual( nested_simplify(__lowerCAmelCase ) , [ [ {'''score''': 0.9_9_9, '''label''': '''Sound of a dog'''}, {'''score''': 0.0_0_1, '''label''': '''Sound of vaccum cleaner'''}, ], ] * 5 , ) @unittest.skip('''No models are available in TF''' ) def _lowerCAmelCase( self ) -> Union[str, Any]: pass
214
'''simple docstring''' import json import os import unittest from transformers import OpenAIGPTTokenizer, OpenAIGPTTokenizerFast from transformers.models.openai.tokenization_openai import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_spacy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class UpperCAmelCase ( a__ , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = OpenAIGPTTokenizer SCREAMING_SNAKE_CASE = OpenAIGPTTokenizerFast SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = False def _lowerCAmelCase( self ) -> Optional[int]: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowercase__ : Optional[Any] = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''w</w>''', '''r</w>''', '''t</w>''', '''lo''', '''low''', '''er</w>''', '''low</w>''', '''lowest</w>''', '''newer</w>''', '''wider</w>''', '''<unk>''', ] lowercase__ : Optional[int] = dict(zip(__lowerCAmelCase , range(len(__lowerCAmelCase ) ) ) ) lowercase__ : Union[str, Any] = ['''#version: 0.2''', '''l o''', '''lo w''', '''e r</w>''', ''''''] lowercase__ : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) lowercase__ : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' ) as fp: fp.write(json.dumps(__lowerCAmelCase ) ) with open(self.merges_file , '''w''' ) as fp: fp.write('''\n'''.join(__lowerCAmelCase ) ) def _lowerCAmelCase( self , __lowerCAmelCase ) -> Dict: return "lower newer", "lower newer" def _lowerCAmelCase( self ) -> List[Any]: lowercase__ : int = OpenAIGPTTokenizer(self.vocab_file , self.merges_file ) lowercase__ : List[Any] = '''lower''' lowercase__ : Any = ['''low''', '''er</w>'''] lowercase__ : Union[str, Any] = tokenizer.tokenize(__lowerCAmelCase ) self.assertListEqual(__lowerCAmelCase , __lowerCAmelCase ) lowercase__ : Tuple = tokens + ['''<unk>'''] lowercase__ : Optional[Any] = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCAmelCase ) , __lowerCAmelCase ) def _lowerCAmelCase( self , __lowerCAmelCase=15 ) -> Optional[int]: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowercase__ : List[str] = self.rust_tokenizer_class.from_pretrained(__lowerCAmelCase , **__lowerCAmelCase ) # Simple input lowercase__ : List[str] = '''This is a simple input''' lowercase__ : Any = ['''This is a simple input 1''', '''This is a simple input 2'''] lowercase__ : str = ('''This is a simple input''', '''This is a pair''') lowercase__ : str = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(__lowerCAmelCase , tokenizer_r.encode , __lowerCAmelCase , max_length=__lowerCAmelCase , padding='''max_length''' ) # Simple input self.assertRaises(__lowerCAmelCase , tokenizer_r.encode_plus , __lowerCAmelCase , max_length=__lowerCAmelCase , padding='''max_length''' ) # Simple input self.assertRaises( __lowerCAmelCase , tokenizer_r.batch_encode_plus , __lowerCAmelCase , max_length=__lowerCAmelCase , padding='''max_length''' , ) # Pair input self.assertRaises(__lowerCAmelCase , tokenizer_r.encode , __lowerCAmelCase , max_length=__lowerCAmelCase , padding='''max_length''' ) # Pair input self.assertRaises(__lowerCAmelCase , tokenizer_r.encode_plus , __lowerCAmelCase , max_length=__lowerCAmelCase , padding='''max_length''' ) # Pair input self.assertRaises( __lowerCAmelCase , tokenizer_r.batch_encode_plus , __lowerCAmelCase , max_length=__lowerCAmelCase , padding='''max_length''' , ) def _lowerCAmelCase( self ) -> Tuple: pass @require_ftfy @require_spacy @require_tokenizers class UpperCAmelCase ( a__ ): '''simple docstring''' pass
214
1
# Lint as: python3 import os import re import urllib.parse from pathlib import Path from typing import Callable, List, Optional, Union from zipfile import ZipFile from ..utils.file_utils import cached_path, hf_github_url from ..utils.logging import get_logger from ..utils.version import Version A__ : Tuple = get_logger(__name__) class __snake_case : _a = '''dummy_data''' _a = '''datasets''' _a = False def __init__( self : Optional[Any] , A_ : str , A_ : str , A_ : Union[Version, str] , A_ : Optional[str] = None , A_ : bool = False , A_ : bool = True , A_ : Optional[List[Callable]] = None , ): lowerCAmelCase_ : List[Any] = 0 lowerCAmelCase_ : Any = dataset_name lowerCAmelCase_ : Union[str, Any] = cache_dir lowerCAmelCase_ : List[Any] = use_local_dummy_data lowerCAmelCase_ : Optional[Any] = config # download_callbacks take a single url as input lowerCAmelCase_ : List[Callable] = download_callbacks or [] # if False, it doesn't load existing files and it returns the paths of the dummy files relative # to the dummy_data zip file root lowerCAmelCase_ : Tuple = load_existing_dummy_data # TODO(PVP, QL) might need to make this more general lowerCAmelCase_ : int = str(A_) # to be downloaded lowerCAmelCase_ : Dict = None lowerCAmelCase_ : Optional[int] = None @property def UpperCAmelCase__ ( self : List[str]): if self._dummy_file is None: lowerCAmelCase_ : int = self.download_dummy_data() return self._dummy_file @property def UpperCAmelCase__ ( self : str): if self.config is not None: # structure is dummy / config_name / version_name return os.path.join('''dummy''' , self.config.name , self.version_name) # structure is dummy / version_name return os.path.join('''dummy''' , self.version_name) @property def UpperCAmelCase__ ( self : str): return os.path.join(self.dummy_data_folder , '''dummy_data.zip''') def UpperCAmelCase__ ( self : Any): lowerCAmelCase_ : Any = ( self.local_path_to_dummy_data if self.use_local_dummy_data is True else self.github_path_to_dummy_data ) lowerCAmelCase_ : Union[str, Any] = cached_path( A_ , cache_dir=self.cache_dir , extract_compressed_file=A_ , force_extract=A_) return os.path.join(A_ , self.dummy_file_name) @property def UpperCAmelCase__ ( self : List[str]): return os.path.join(self.datasets_scripts_dir , self.dataset_name , self.dummy_zip_file) @property def UpperCAmelCase__ ( self : Optional[int]): if self._bucket_url is None: lowerCAmelCase_ : str = hf_github_url(self.dataset_name , self.dummy_zip_file.replace(os.sep , '''/''')) return self._bucket_url @property def UpperCAmelCase__ ( self : List[Any]): # return full path if its a dir if os.path.isdir(self.dummy_file): return self.dummy_file # else cut off path to file -> example `xsum`. return "/".join(self.dummy_file.replace(os.sep , '''/''').split('''/''')[:-1]) def UpperCAmelCase__ ( self : Union[str, Any] , A_ : Dict , *A_ : List[Any]): if self.load_existing_dummy_data: # dummy data is downloaded and tested lowerCAmelCase_ : Union[str, Any] = self.dummy_file else: # dummy data cannot be downloaded and only the path to dummy file is returned lowerCAmelCase_ : Optional[int] = self.dummy_file_name # special case when data_url is a dict if isinstance(A_ , A_): return self.create_dummy_data_dict(A_ , A_) elif isinstance(A_ , (list, tuple)): return self.create_dummy_data_list(A_ , A_) else: return self.create_dummy_data_single(A_ , A_) def UpperCAmelCase__ ( self : Optional[int] , A_ : Tuple , *A_ : int): return self.download_and_extract(A_) def UpperCAmelCase__ ( self : Tuple , A_ : List[str] , A_ : Optional[Any]): return self.download_and_extract(A_) def UpperCAmelCase__ ( self : int , A_ : Optional[int] , *A_ : str , **A_ : List[Any]): return path def UpperCAmelCase__ ( self : Tuple): return {} def UpperCAmelCase__ ( self : Optional[Any] , A_ : Union[str, Any] , A_ : List[Any]): lowerCAmelCase_ : Union[str, Any] = {} for key, single_urls in data_url.items(): for download_callback in self.download_callbacks: if isinstance(A_ , A_): for single_url in single_urls: download_callback(A_) else: lowerCAmelCase_ : Any = single_urls download_callback(A_) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus if isinstance(A_ , A_): lowerCAmelCase_ : Any = [os.path.join(A_ , urllib.parse.quote_plus(Path(A_).name)) for x in single_urls] else: lowerCAmelCase_ : Optional[int] = single_urls lowerCAmelCase_ : List[str] = os.path.join(A_ , urllib.parse.quote_plus(Path(A_).name)) lowerCAmelCase_ : Dict = value # make sure that values are unique if all(isinstance(A_ , A_) for i in dummy_data_dict.values()) and len(set(dummy_data_dict.values())) < len( dummy_data_dict.values()): # append key to value to make its name unique lowerCAmelCase_ : Tuple = {key: value + key for key, value in dummy_data_dict.items()} return dummy_data_dict def UpperCAmelCase__ ( self : Dict , A_ : List[str] , A_ : str): lowerCAmelCase_ : Optional[Any] = [] # trick: if there are many shards named like `data.txt-000001-of-00300`, only use the first one lowerCAmelCase_ : str = all(bool(re.findall('''[0-9]{3,}-of-[0-9]{3,}''' , A_)) for url in data_url) lowerCAmelCase_ : Optional[Any] = all( url.startswith('''https://ftp.ncbi.nlm.nih.gov/pubmed/baseline/pubmed''') for url in data_url) if data_url and (is_tf_records or is_pubmed_records): lowerCAmelCase_ : Any = [data_url[0]] * len(A_) for single_url in data_url: for download_callback in self.download_callbacks: download_callback(A_) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus lowerCAmelCase_ : int = os.path.join(A_ , urllib.parse.quote_plus(single_url.split('''/''')[-1])) dummy_data_list.append(A_) return dummy_data_list def UpperCAmelCase__ ( self : List[str] , A_ : Optional[Any] , A_ : Tuple): for download_callback in self.download_callbacks: download_callback(A_) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus lowerCAmelCase_ : Tuple = os.path.join(A_ , urllib.parse.quote_plus(data_url.split('''/''')[-1])) if os.path.exists(A_) or not self.load_existing_dummy_data: return value else: # Backward compatibility, maybe deprecate at one point. # For many datasets with single url calls to dl_manager.download_and_extract, # the dummy_data.zip file is actually the zipped downloaded file # while now we expected the dummy_data.zip file to be a directory containing # the downloaded file. return path_to_dummy_data def UpperCAmelCase__ ( self : int): pass def UpperCAmelCase__ ( self : Optional[int]): pass def UpperCAmelCase__ ( self : List[str] , A_ : str): def _iter_archive_members(A_ : Any): # this preserves the order of the members inside the ZIP archive lowerCAmelCase_ : Optional[int] = Path(self.dummy_file).parent lowerCAmelCase_ : Optional[int] = path.relative_to(A_) with ZipFile(self.local_path_to_dummy_data) as zip_file: lowerCAmelCase_ : Tuple = zip_file.namelist() for member in members: if member.startswith(relative_path.as_posix()): yield dummy_parent_path.joinpath(A_) lowerCAmelCase_ : List[Any] = Path(A_) lowerCAmelCase_ : Optional[int] = _iter_archive_members(A_) if self.use_local_dummy_data else path.rglob('''*''') for file_path in file_paths: if file_path.is_file() and not file_path.name.startswith(('''.''', '''__''')): yield file_path.relative_to(A_).as_posix(), file_path.open('''rb''') def UpperCAmelCase__ ( self : Dict , A_ : Any): if not isinstance(A_ , A_): lowerCAmelCase_ : Dict = [paths] for path in paths: if os.path.isfile(A_): if os.path.basename(A_).startswith(('''.''', '''__''')): return yield path else: for dirpath, dirnames, filenames in os.walk(A_): if os.path.basename(A_).startswith(('''.''', '''__''')): continue dirnames.sort() for filename in sorted(A_): if filename.startswith(('''.''', '''__''')): continue yield os.path.join(A_ , A_)
103
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() A__ : List[Any] = logging.get_logger(__name__) def UpperCamelCase( __UpperCamelCase : List[str] ): lowerCAmelCase_ : Any = DPTConfig(embedding_type='''hybrid''' ) if "large" in checkpoint_url: lowerCAmelCase_ : Any = 1024 lowerCAmelCase_ : Tuple = 4096 lowerCAmelCase_ : List[Any] = 24 lowerCAmelCase_ : int = 16 lowerCAmelCase_ : Tuple = [5, 11, 17, 23] lowerCAmelCase_ : Optional[Any] = [256, 512, 1024, 1024] lowerCAmelCase_ : str = (1, 384, 384) if "nyu" or "midas" in checkpoint_url: lowerCAmelCase_ : Union[str, Any] = 768 lowerCAmelCase_ : int = [1, 1, 1, 0.5] lowerCAmelCase_ : List[str] = [256, 512, 768, 768] lowerCAmelCase_ : int = 150 lowerCAmelCase_ : Any = 16 lowerCAmelCase_ : Any = (1, 384, 384) lowerCAmelCase_ : List[str] = False lowerCAmelCase_ : Union[str, Any] = '''project''' if "ade" in checkpoint_url: lowerCAmelCase_ : Any = True lowerCAmelCase_ : Any = 768 lowerCAmelCase_ : Optional[int] = [1, 1, 1, 0.5] lowerCAmelCase_ : Union[str, Any] = 150 lowerCAmelCase_ : Any = 16 lowerCAmelCase_ : Any = '''huggingface/label-files''' lowerCAmelCase_ : Any = '''ade20k-id2label.json''' lowerCAmelCase_ : str = json.load(open(cached_download(hf_hub_url(__UpperCamelCase ,__UpperCamelCase ,repo_type='''dataset''' ) ) ,'''r''' ) ) lowerCAmelCase_ : Optional[Any] = {int(__UpperCamelCase ): v for k, v in idalabel.items()} lowerCAmelCase_ : Optional[Any] = idalabel lowerCAmelCase_ : Optional[int] = {v: k for k, v in idalabel.items()} lowerCAmelCase_ : List[Any] = [1, 150, 480, 480] return config, expected_shape def UpperCamelCase( __UpperCamelCase : Optional[int] ): lowerCAmelCase_ : int = ['''pretrained.model.head.weight''', '''pretrained.model.head.bias'''] for k in ignore_keys: state_dict.pop(__UpperCamelCase ,__UpperCamelCase ) def UpperCamelCase( __UpperCamelCase : int ): if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): lowerCAmelCase_ : Dict = name.replace('''pretrained.model''' ,'''dpt.encoder''' ) if "pretrained.model" in name: lowerCAmelCase_ : List[Any] = name.replace('''pretrained.model''' ,'''dpt.embeddings''' ) if "patch_embed" in name: lowerCAmelCase_ : Optional[Any] = name.replace('''patch_embed''' ,'''''' ) if "pos_embed" in name: lowerCAmelCase_ : Dict = name.replace('''pos_embed''' ,'''position_embeddings''' ) if "attn.proj" in name: lowerCAmelCase_ : Any = name.replace('''attn.proj''' ,'''attention.output.dense''' ) if "proj" in name and "project" not in name: lowerCAmelCase_ : Tuple = name.replace('''proj''' ,'''projection''' ) if "blocks" in name: lowerCAmelCase_ : Optional[Any] = name.replace('''blocks''' ,'''layer''' ) if "mlp.fc1" in name: lowerCAmelCase_ : Optional[int] = name.replace('''mlp.fc1''' ,'''intermediate.dense''' ) if "mlp.fc2" in name: lowerCAmelCase_ : Union[str, Any] = name.replace('''mlp.fc2''' ,'''output.dense''' ) if "norm1" in name and "backbone" not in name: lowerCAmelCase_ : List[str] = name.replace('''norm1''' ,'''layernorm_before''' ) if "norm2" in name and "backbone" not in name: lowerCAmelCase_ : Optional[Any] = name.replace('''norm2''' ,'''layernorm_after''' ) if "scratch.output_conv" in name: lowerCAmelCase_ : Optional[int] = name.replace('''scratch.output_conv''' ,'''head''' ) if "scratch" in name: lowerCAmelCase_ : Dict = name.replace('''scratch''' ,'''neck''' ) if "layer1_rn" in name: lowerCAmelCase_ : Optional[int] = name.replace('''layer1_rn''' ,'''convs.0''' ) if "layer2_rn" in name: lowerCAmelCase_ : Union[str, Any] = name.replace('''layer2_rn''' ,'''convs.1''' ) if "layer3_rn" in name: lowerCAmelCase_ : List[Any] = name.replace('''layer3_rn''' ,'''convs.2''' ) if "layer4_rn" in name: lowerCAmelCase_ : Optional[int] = name.replace('''layer4_rn''' ,'''convs.3''' ) if "refinenet" in name: lowerCAmelCase_ : List[str] = int(name[len('''neck.refinenet''' ) : len('''neck.refinenet''' ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 lowerCAmelCase_ : Dict = name.replace(f"""refinenet{layer_idx}""" ,f"""fusion_stage.layers.{abs(layer_idx-4 )}""" ) if "out_conv" in name: lowerCAmelCase_ : int = name.replace('''out_conv''' ,'''projection''' ) if "resConfUnit1" in name: lowerCAmelCase_ : Dict = name.replace('''resConfUnit1''' ,'''residual_layer1''' ) if "resConfUnit2" in name: lowerCAmelCase_ : str = name.replace('''resConfUnit2''' ,'''residual_layer2''' ) if "conv1" in name: lowerCAmelCase_ : str = name.replace('''conv1''' ,'''convolution1''' ) if "conv2" in name: lowerCAmelCase_ : Optional[Any] = name.replace('''conv2''' ,'''convolution2''' ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: lowerCAmelCase_ : Dict = name.replace('''pretrained.act_postprocess1.0.project.0''' ,'''neck.reassemble_stage.readout_projects.0.0''' ) if "pretrained.act_postprocess2.0.project.0" in name: lowerCAmelCase_ : Optional[int] = name.replace('''pretrained.act_postprocess2.0.project.0''' ,'''neck.reassemble_stage.readout_projects.1.0''' ) if "pretrained.act_postprocess3.0.project.0" in name: lowerCAmelCase_ : Tuple = name.replace('''pretrained.act_postprocess3.0.project.0''' ,'''neck.reassemble_stage.readout_projects.2.0''' ) if "pretrained.act_postprocess4.0.project.0" in name: lowerCAmelCase_ : Dict = name.replace('''pretrained.act_postprocess4.0.project.0''' ,'''neck.reassemble_stage.readout_projects.3.0''' ) # resize blocks if "pretrained.act_postprocess1.3" in name: lowerCAmelCase_ : List[Any] = name.replace('''pretrained.act_postprocess1.3''' ,'''neck.reassemble_stage.layers.0.projection''' ) if "pretrained.act_postprocess1.4" in name: lowerCAmelCase_ : Dict = name.replace('''pretrained.act_postprocess1.4''' ,'''neck.reassemble_stage.layers.0.resize''' ) if "pretrained.act_postprocess2.3" in name: lowerCAmelCase_ : List[Any] = name.replace('''pretrained.act_postprocess2.3''' ,'''neck.reassemble_stage.layers.1.projection''' ) if "pretrained.act_postprocess2.4" in name: lowerCAmelCase_ : List[str] = name.replace('''pretrained.act_postprocess2.4''' ,'''neck.reassemble_stage.layers.1.resize''' ) if "pretrained.act_postprocess3.3" in name: lowerCAmelCase_ : Optional[Any] = name.replace('''pretrained.act_postprocess3.3''' ,'''neck.reassemble_stage.layers.2.projection''' ) if "pretrained.act_postprocess4.3" in name: lowerCAmelCase_ : List[str] = name.replace('''pretrained.act_postprocess4.3''' ,'''neck.reassemble_stage.layers.3.projection''' ) if "pretrained.act_postprocess4.4" in name: lowerCAmelCase_ : Optional[Any] = name.replace('''pretrained.act_postprocess4.4''' ,'''neck.reassemble_stage.layers.3.resize''' ) if "pretrained" in name: lowerCAmelCase_ : Tuple = name.replace('''pretrained''' ,'''dpt''' ) if "bn" in name: lowerCAmelCase_ : Dict = name.replace('''bn''' ,'''batch_norm''' ) if "head" in name: lowerCAmelCase_ : Any = name.replace('''head''' ,'''head.head''' ) if "encoder.norm" in name: lowerCAmelCase_ : Tuple = name.replace('''encoder.norm''' ,'''layernorm''' ) if "auxlayer" in name: lowerCAmelCase_ : Optional[int] = name.replace('''auxlayer''' ,'''auxiliary_head.head''' ) if "backbone" in name: lowerCAmelCase_ : List[Any] = name.replace('''backbone''' ,'''backbone.bit.encoder''' ) if ".." in name: lowerCAmelCase_ : List[Any] = name.replace('''..''' ,'''.''' ) if "stem.conv" in name: lowerCAmelCase_ : str = name.replace('''stem.conv''' ,'''bit.embedder.convolution''' ) if "blocks" in name: lowerCAmelCase_ : List[str] = name.replace('''blocks''' ,'''layers''' ) if "convolution" in name and "backbone" in name: lowerCAmelCase_ : Optional[int] = name.replace('''convolution''' ,'''conv''' ) if "layer" in name and "backbone" in name: lowerCAmelCase_ : Optional[int] = name.replace('''layer''' ,'''layers''' ) if "backbone.bit.encoder.bit" in name: lowerCAmelCase_ : Union[str, Any] = name.replace('''backbone.bit.encoder.bit''' ,'''backbone.bit''' ) if "embedder.conv" in name: lowerCAmelCase_ : str = name.replace('''embedder.conv''' ,'''embedder.convolution''' ) if "backbone.bit.encoder.stem.norm" in name: lowerCAmelCase_ : Dict = name.replace('''backbone.bit.encoder.stem.norm''' ,'''backbone.bit.embedder.norm''' ) return name def UpperCamelCase( __UpperCamelCase : Union[str, Any] ,__UpperCamelCase : List[str] ): for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowerCAmelCase_ : Dict = state_dict.pop(f"""dpt.encoder.layer.{i}.attn.qkv.weight""" ) lowerCAmelCase_ : Tuple = state_dict.pop(f"""dpt.encoder.layer.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict lowerCAmelCase_ : str = in_proj_weight[: config.hidden_size, :] lowerCAmelCase_ : str = in_proj_bias[: config.hidden_size] lowerCAmelCase_ : Optional[Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowerCAmelCase_ : int = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowerCAmelCase_ : Any = in_proj_weight[ -config.hidden_size :, : ] lowerCAmelCase_ : Dict = in_proj_bias[-config.hidden_size :] def UpperCamelCase( ): lowerCAmelCase_ : str = '''http://images.cocodataset.org/val2017/000000039769.jpg''' lowerCAmelCase_ : Dict = Image.open(requests.get(__UpperCamelCase ,stream=__UpperCamelCase ).raw ) return im @torch.no_grad() def UpperCamelCase( __UpperCamelCase : Optional[int] ,__UpperCamelCase : Dict ,__UpperCamelCase : Optional[int] ,__UpperCamelCase : Optional[int] ,__UpperCamelCase : str ): lowerCAmelCase_ , lowerCAmelCase_ : Any = get_dpt_config(__UpperCamelCase ) # load original state_dict from URL # state_dict = torch.hub.load_state_dict_from_url(checkpoint_url, map_location="cpu") lowerCAmelCase_ : List[str] = torch.load(__UpperCamelCase ,map_location='''cpu''' ) # remove certain keys remove_ignore_keys_(__UpperCamelCase ) # rename keys for key in state_dict.copy().keys(): lowerCAmelCase_ : Any = state_dict.pop(__UpperCamelCase ) lowerCAmelCase_ : Optional[Any] = val # read in qkv matrices read_in_q_k_v(__UpperCamelCase ,__UpperCamelCase ) # load HuggingFace model lowerCAmelCase_ : List[Any] = DPTForSemanticSegmentation(__UpperCamelCase ) if '''ade''' in checkpoint_url else DPTForDepthEstimation(__UpperCamelCase ) model.load_state_dict(__UpperCamelCase ) model.eval() # Check outputs on an image lowerCAmelCase_ : Tuple = 480 if '''ade''' in checkpoint_url else 384 lowerCAmelCase_ : Optional[int] = DPTImageProcessor(size=__UpperCamelCase ) lowerCAmelCase_ : Union[str, Any] = prepare_img() lowerCAmelCase_ : str = image_processor(__UpperCamelCase ,return_tensors='''pt''' ) # forward pass lowerCAmelCase_ : Tuple = model(**__UpperCamelCase ).logits if '''ade''' in checkpoint_url else model(**__UpperCamelCase ).predicted_depth if show_prediction: lowerCAmelCase_ : Optional[Any] = ( torch.nn.functional.interpolate( outputs.unsqueeze(1 ) ,size=(image.size[1], image.size[0]) ,mode='''bicubic''' ,align_corners=__UpperCamelCase ,) .squeeze() .cpu() .numpy() ) Image.fromarray((prediction / prediction.max()) * 255 ).show() if pytorch_dump_folder_path is not None: Path(__UpperCamelCase ).mkdir(exist_ok=__UpperCamelCase ) print(f"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(__UpperCamelCase ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__UpperCamelCase ) if push_to_hub: model.push_to_hub('''ybelkada/dpt-hybrid-midas''' ) image_processor.push_to_hub('''ybelkada/dpt-hybrid-midas''' ) if __name__ == "__main__": A__ : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt''', type=str, help='''URL of the original DPT checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=False, help='''Path to the output PyTorch model directory.''', ) parser.add_argument( '''--push_to_hub''', action='''store_true''', ) parser.add_argument( '''--model_name''', default='''dpt-large''', type=str, help='''Name of the model, in case you\'re pushing to the hub.''', ) parser.add_argument( '''--show_prediction''', action='''store_true''', ) A__ : Union[str, Any] = parser.parse_args() convert_dpt_checkpoint( args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name, args.show_prediction )
103
1
"""simple docstring""" import os import unittest from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, BertTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class lowercase__ ( SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' UpperCamelCase = BertTokenizer UpperCamelCase = BertTokenizerFast UpperCamelCase = True UpperCamelCase = True UpperCamelCase = filter_non_english def lowercase__ ( self : int ) -> Any: '''simple docstring''' super().setUp() UpperCAmelCase_ = [ "[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] UpperCAmelCase_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def lowercase__ ( self : int , _UpperCAmelCase : List[Any] ) -> List[str]: '''simple docstring''' UpperCAmelCase_ = "UNwant\u00E9d,running" UpperCAmelCase_ = "unwanted, running" return input_text, output_text def lowercase__ ( self : int ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ = self.tokenizer_class(self.vocab_file ) UpperCAmelCase_ = tokenizer.tokenize("UNwant\u00E9d,running" ) self.assertListEqual(_UpperCAmelCase , ["un", "##want", "##ed", ",", "runn", "##ing"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_UpperCAmelCase ) , [9, 6, 7, 12, 10, 11] ) def lowercase__ ( self : Union[str, Any] ) -> Any: '''simple docstring''' if not self.test_rust_tokenizer: return UpperCAmelCase_ = self.get_tokenizer() UpperCAmelCase_ = self.get_rust_tokenizer() UpperCAmelCase_ = "UNwant\u00E9d,running" UpperCAmelCase_ = tokenizer.tokenize(_UpperCAmelCase ) UpperCAmelCase_ = rust_tokenizer.tokenize(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) UpperCAmelCase_ = tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) UpperCAmelCase_ = rust_tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) UpperCAmelCase_ = self.get_rust_tokenizer() UpperCAmelCase_ = tokenizer.encode(_UpperCAmelCase ) UpperCAmelCase_ = rust_tokenizer.encode(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) # With lower casing UpperCAmelCase_ = self.get_tokenizer(do_lower_case=_UpperCAmelCase ) UpperCAmelCase_ = self.get_rust_tokenizer(do_lower_case=_UpperCAmelCase ) UpperCAmelCase_ = "UNwant\u00E9d,running" UpperCAmelCase_ = tokenizer.tokenize(_UpperCAmelCase ) UpperCAmelCase_ = rust_tokenizer.tokenize(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) UpperCAmelCase_ = tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) UpperCAmelCase_ = rust_tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) UpperCAmelCase_ = self.get_rust_tokenizer() UpperCAmelCase_ = tokenizer.encode(_UpperCAmelCase ) UpperCAmelCase_ = rust_tokenizer.encode(_UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) def lowercase__ ( self : List[Any] ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase_ = BasicTokenizer() self.assertListEqual(tokenizer.tokenize("ah\u535A\u63A8zz" ) , ["ah", "\u535A", "\u63A8", "zz"] ) def lowercase__ ( self : Optional[Any] ) -> str: '''simple docstring''' UpperCAmelCase_ = BasicTokenizer(do_lower_case=_UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? " ) , ["hello", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["hello"] ) def lowercase__ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase_ = BasicTokenizer(do_lower_case=_UpperCAmelCase , strip_accents=_UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["hällo", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["h\u00E9llo"] ) def lowercase__ ( self : List[Any] ) -> Dict: '''simple docstring''' UpperCAmelCase_ = BasicTokenizer(do_lower_case=_UpperCAmelCase , strip_accents=_UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["hallo", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["hello"] ) def lowercase__ ( self : int ) -> int: '''simple docstring''' UpperCAmelCase_ = BasicTokenizer(do_lower_case=_UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["hallo", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["hello"] ) def lowercase__ ( self : Dict ) -> Dict: '''simple docstring''' UpperCAmelCase_ = BasicTokenizer(do_lower_case=_UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? " ) , ["HeLLo", "!", "how", "Are", "yoU", "?"] ) def lowercase__ ( self : Union[str, Any] ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase_ = BasicTokenizer(do_lower_case=_UpperCAmelCase , strip_accents=_UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["HäLLo", "!", "how", "Are", "yoU", "?"] ) def lowercase__ ( self : int ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase_ = BasicTokenizer(do_lower_case=_UpperCAmelCase , strip_accents=_UpperCAmelCase ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["HaLLo", "!", "how", "Are", "yoU", "?"] ) def lowercase__ ( self : Optional[Any] ) -> Tuple: '''simple docstring''' UpperCAmelCase_ = BasicTokenizer(do_lower_case=_UpperCAmelCase , never_split=["[UNK]"] ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? [UNK]" ) , ["HeLLo", "!", "how", "Are", "yoU", "?", "[UNK]"] ) def lowercase__ ( self : Any ) -> Any: '''simple docstring''' UpperCAmelCase_ = BasicTokenizer() UpperCAmelCase_ = "a\n'll !!to?'d of, can't." UpperCAmelCase_ = ["a", "'", "ll", "!", "!", "to", "?", "'", "d", "of", ",", "can", "'", "t", "."] self.assertListEqual(tokenizer.tokenize(_UpperCAmelCase ) , _UpperCAmelCase ) def lowercase__ ( self : Optional[int] ) -> str: '''simple docstring''' UpperCAmelCase_ = ["[UNK]", "[CLS]", "[SEP]", "want", "##want", "##ed", "wa", "un", "runn", "##ing"] UpperCAmelCase_ = {} for i, token in enumerate(_UpperCAmelCase ): UpperCAmelCase_ = i UpperCAmelCase_ = WordpieceTokenizer(vocab=_UpperCAmelCase , unk_token="[UNK]" ) self.assertListEqual(tokenizer.tokenize("" ) , [] ) self.assertListEqual(tokenizer.tokenize("unwanted running" ) , ["un", "##want", "##ed", "runn", "##ing"] ) self.assertListEqual(tokenizer.tokenize("unwantedX running" ) , ["[UNK]", "runn", "##ing"] ) def lowercase__ ( self : Tuple ) -> List[str]: '''simple docstring''' self.assertTrue(_is_whitespace(" " ) ) self.assertTrue(_is_whitespace("\t" ) ) self.assertTrue(_is_whitespace("\r" ) ) self.assertTrue(_is_whitespace("\n" ) ) self.assertTrue(_is_whitespace("\u00A0" ) ) self.assertFalse(_is_whitespace("A" ) ) self.assertFalse(_is_whitespace("-" ) ) def lowercase__ ( self : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' self.assertTrue(_is_control("\u0005" ) ) self.assertFalse(_is_control("A" ) ) self.assertFalse(_is_control(" " ) ) self.assertFalse(_is_control("\t" ) ) self.assertFalse(_is_control("\r" ) ) def lowercase__ ( self : Optional[Any] ) -> List[str]: '''simple docstring''' self.assertTrue(_is_punctuation("-" ) ) self.assertTrue(_is_punctuation("$" ) ) self.assertTrue(_is_punctuation("`" ) ) self.assertTrue(_is_punctuation("." ) ) self.assertFalse(_is_punctuation("A" ) ) self.assertFalse(_is_punctuation(" " ) ) def lowercase__ ( self : int ) -> Tuple: '''simple docstring''' UpperCAmelCase_ = self.get_tokenizer() UpperCAmelCase_ = self.get_rust_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(_UpperCAmelCase ) for t in ["Test", "\xad", "test"]] , [["[UNK]"], [], ["[UNK]"]] ) self.assertListEqual( [rust_tokenizer.tokenize(_UpperCAmelCase ) for t in ["Test", "\xad", "test"]] , [["[UNK]"], [], ["[UNK]"]] ) @slow def lowercase__ ( self : Any ) -> List[Any]: '''simple docstring''' UpperCAmelCase_ = self.tokenizer_class.from_pretrained("bert-base-uncased" ) UpperCAmelCase_ = tokenizer.encode("sequence builders" , add_special_tokens=_UpperCAmelCase ) UpperCAmelCase_ = tokenizer.encode("multi-sequence build" , add_special_tokens=_UpperCAmelCase ) UpperCAmelCase_ = tokenizer.build_inputs_with_special_tokens(_UpperCAmelCase ) UpperCAmelCase_ = tokenizer.build_inputs_with_special_tokens(_UpperCAmelCase , _UpperCAmelCase ) assert encoded_sentence == [101] + text + [102] assert encoded_pair == [101] + text + [102] + text_a + [102] def lowercase__ ( self : Optional[int] ) -> str: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): UpperCAmelCase_ = self.rust_tokenizer_class.from_pretrained(_UpperCAmelCase , **_UpperCAmelCase ) UpperCAmelCase_ = F"""A, naïve {tokenizer_r.mask_token} AllenNLP sentence.""" UpperCAmelCase_ = tokenizer_r.encode_plus( _UpperCAmelCase , return_attention_mask=_UpperCAmelCase , return_token_type_ids=_UpperCAmelCase , return_offsets_mapping=_UpperCAmelCase , add_special_tokens=_UpperCAmelCase , ) UpperCAmelCase_ = tokenizer_r.do_lower_case if hasattr(_UpperCAmelCase , "do_lower_case" ) else False UpperCAmelCase_ = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), "A"), ((1, 2), ","), ((3, 5), "na"), ((5, 6), "##ï"), ((6, 8), "##ve"), ((9, 15), tokenizer_r.mask_token), ((16, 21), "Allen"), ((21, 23), "##NL"), ((23, 24), "##P"), ((25, 33), "sentence"), ((33, 34), "."), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), "a"), ((1, 2), ","), ((3, 8), "naive"), ((9, 15), tokenizer_r.mask_token), ((16, 21), "allen"), ((21, 23), "##nl"), ((23, 24), "##p"), ((25, 33), "sentence"), ((33, 34), "."), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens["input_ids"] ) ) self.assertEqual([e[0] for e in expected_results] , tokens["offset_mapping"] ) def lowercase__ ( self : Union[str, Any] ) -> Any: '''simple docstring''' UpperCAmelCase_ = ["的", "人", "有"] UpperCAmelCase_ = "".join(_UpperCAmelCase ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): UpperCAmelCase_ = True UpperCAmelCase_ = self.tokenizer_class.from_pretrained(_UpperCAmelCase , **_UpperCAmelCase ) UpperCAmelCase_ = self.rust_tokenizer_class.from_pretrained(_UpperCAmelCase , **_UpperCAmelCase ) UpperCAmelCase_ = tokenizer_p.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) UpperCAmelCase_ = tokenizer_r.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) UpperCAmelCase_ = tokenizer_r.convert_ids_to_tokens(_UpperCAmelCase ) UpperCAmelCase_ = tokenizer_p.convert_ids_to_tokens(_UpperCAmelCase ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) UpperCAmelCase_ = False UpperCAmelCase_ = self.rust_tokenizer_class.from_pretrained(_UpperCAmelCase , **_UpperCAmelCase ) UpperCAmelCase_ = self.tokenizer_class.from_pretrained(_UpperCAmelCase , **_UpperCAmelCase ) UpperCAmelCase_ = tokenizer_r.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) UpperCAmelCase_ = tokenizer_p.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ) UpperCAmelCase_ = tokenizer_r.convert_ids_to_tokens(_UpperCAmelCase ) UpperCAmelCase_ = tokenizer_p.convert_ids_to_tokens(_UpperCAmelCase ) # it is expected that only the first Chinese character is not preceded by "##". UpperCAmelCase_ = [ F"""##{token}""" if idx != 0 else token for idx, token in enumerate(_UpperCAmelCase ) ] self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase )
241
"""simple docstring""" import argparse import io import requests import torch from omegaconf import OmegaConf from diffusers import AutoencoderKL from diffusers.pipelines.stable_diffusion.convert_from_ckpt import ( assign_to_checkpoint, conv_attn_to_linear, create_vae_diffusers_config, renew_vae_attention_paths, renew_vae_resnet_paths, ) def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = checkpoint UpperCAmelCase_ = {} UpperCAmelCase_ = vae_state_dict["encoder.conv_in.weight"] UpperCAmelCase_ = vae_state_dict["encoder.conv_in.bias"] UpperCAmelCase_ = vae_state_dict["encoder.conv_out.weight"] UpperCAmelCase_ = vae_state_dict["encoder.conv_out.bias"] UpperCAmelCase_ = vae_state_dict["encoder.norm_out.weight"] UpperCAmelCase_ = vae_state_dict["encoder.norm_out.bias"] UpperCAmelCase_ = vae_state_dict["decoder.conv_in.weight"] UpperCAmelCase_ = vae_state_dict["decoder.conv_in.bias"] UpperCAmelCase_ = vae_state_dict["decoder.conv_out.weight"] UpperCAmelCase_ = vae_state_dict["decoder.conv_out.bias"] UpperCAmelCase_ = vae_state_dict["decoder.norm_out.weight"] UpperCAmelCase_ = vae_state_dict["decoder.norm_out.bias"] UpperCAmelCase_ = vae_state_dict["quant_conv.weight"] UpperCAmelCase_ = vae_state_dict["quant_conv.bias"] UpperCAmelCase_ = vae_state_dict["post_quant_conv.weight"] UpperCAmelCase_ = vae_state_dict["post_quant_conv.bias"] # Retrieves the keys for the encoder down blocks only UpperCAmelCase_ = len({".".join(layer.split("." )[:3] ) for layer in vae_state_dict if "encoder.down" in layer} ) UpperCAmelCase_ = { layer_id: [key for key in vae_state_dict if f"""down.{layer_id}""" in key] for layer_id in range(lowerCAmelCase__ ) } # Retrieves the keys for the decoder up blocks only UpperCAmelCase_ = len({".".join(layer.split("." )[:3] ) for layer in vae_state_dict if "decoder.up" in layer} ) UpperCAmelCase_ = { layer_id: [key for key in vae_state_dict if f"""up.{layer_id}""" in key] for layer_id in range(lowerCAmelCase__ ) } for i in range(lowerCAmelCase__ ): UpperCAmelCase_ = [key for key in down_blocks[i] if f"""down.{i}""" in key and f"""down.{i}.downsample""" not in key] if f"""encoder.down.{i}.downsample.conv.weight""" in vae_state_dict: UpperCAmelCase_ = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.weight""" ) UpperCAmelCase_ = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.bias""" ) UpperCAmelCase_ = renew_vae_resnet_paths(lowerCAmelCase__ ) UpperCAmelCase_ = {"old": f"""down.{i}.block""", "new": f"""down_blocks.{i}.resnets"""} assign_to_checkpoint(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , additional_replacements=[meta_path] , config=lowerCAmelCase__ ) UpperCAmelCase_ = [key for key in vae_state_dict if "encoder.mid.block" in key] UpperCAmelCase_ = 2 for i in range(1 , num_mid_res_blocks + 1 ): UpperCAmelCase_ = [key for key in mid_resnets if f"""encoder.mid.block_{i}""" in key] UpperCAmelCase_ = renew_vae_resnet_paths(lowerCAmelCase__ ) UpperCAmelCase_ = {"old": f"""mid.block_{i}""", "new": f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , additional_replacements=[meta_path] , config=lowerCAmelCase__ ) UpperCAmelCase_ = [key for key in vae_state_dict if "encoder.mid.attn" in key] UpperCAmelCase_ = renew_vae_attention_paths(lowerCAmelCase__ ) UpperCAmelCase_ = {"old": "mid.attn_1", "new": "mid_block.attentions.0"} assign_to_checkpoint(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , additional_replacements=[meta_path] , config=lowerCAmelCase__ ) conv_attn_to_linear(lowerCAmelCase__ ) for i in range(lowerCAmelCase__ ): UpperCAmelCase_ = num_up_blocks - 1 - i UpperCAmelCase_ = [ key for key in up_blocks[block_id] if f"""up.{block_id}""" in key and f"""up.{block_id}.upsample""" not in key ] if f"""decoder.up.{block_id}.upsample.conv.weight""" in vae_state_dict: UpperCAmelCase_ = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.weight""" ] UpperCAmelCase_ = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.bias""" ] UpperCAmelCase_ = renew_vae_resnet_paths(lowerCAmelCase__ ) UpperCAmelCase_ = {"old": f"""up.{block_id}.block""", "new": f"""up_blocks.{i}.resnets"""} assign_to_checkpoint(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , additional_replacements=[meta_path] , config=lowerCAmelCase__ ) UpperCAmelCase_ = [key for key in vae_state_dict if "decoder.mid.block" in key] UpperCAmelCase_ = 2 for i in range(1 , num_mid_res_blocks + 1 ): UpperCAmelCase_ = [key for key in mid_resnets if f"""decoder.mid.block_{i}""" in key] UpperCAmelCase_ = renew_vae_resnet_paths(lowerCAmelCase__ ) UpperCAmelCase_ = {"old": f"""mid.block_{i}""", "new": f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , additional_replacements=[meta_path] , config=lowerCAmelCase__ ) UpperCAmelCase_ = [key for key in vae_state_dict if "decoder.mid.attn" in key] UpperCAmelCase_ = renew_vae_attention_paths(lowerCAmelCase__ ) UpperCAmelCase_ = {"old": "mid.attn_1", "new": "mid_block.attentions.0"} assign_to_checkpoint(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , additional_replacements=[meta_path] , config=lowerCAmelCase__ ) conv_attn_to_linear(lowerCAmelCase__ ) return new_checkpoint def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , ): # Only support V1 UpperCAmelCase_ = requests.get( " https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml" ) UpperCAmelCase_ = io.BytesIO(r.content ) UpperCAmelCase_ = OmegaConf.load(lowerCAmelCase__ ) UpperCAmelCase_ = 512 UpperCAmelCase_ = "cuda" if torch.cuda.is_available() else "cpu" if checkpoint_path.endswith("safetensors" ): from safetensors import safe_open UpperCAmelCase_ = {} with safe_open(lowerCAmelCase__ , framework="pt" , device="cpu" ) as f: for key in f.keys(): UpperCAmelCase_ = f.get_tensor(lowerCAmelCase__ ) else: UpperCAmelCase_ = torch.load(lowerCAmelCase__ , map_location=lowerCAmelCase__ )["state_dict"] # Convert the VAE model. UpperCAmelCase_ = create_vae_diffusers_config(lowerCAmelCase__ , image_size=lowerCAmelCase__ ) UpperCAmelCase_ = custom_convert_ldm_vae_checkpoint(lowerCAmelCase__ , lowerCAmelCase__ ) UpperCAmelCase_ = AutoencoderKL(**lowerCAmelCase__ ) vae.load_state_dict(lowerCAmelCase__ ) vae.save_pretrained(lowerCAmelCase__ ) if __name__ == "__main__": lowerCamelCase = argparse.ArgumentParser() parser.add_argument("""--vae_pt_path""", default=None, type=str, required=True, help="""Path to the VAE.pt to convert.""") parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the VAE.pt to convert.""") lowerCamelCase = parser.parse_args() vae_pt_to_vae_diffuser(args.vae_pt_path, args.dump_path)
241
1
'''simple docstring''' import warnings from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch from ...models import UNetaDModel from ...schedulers import RePaintScheduler from ...utils import PIL_INTERPOLATION, logging, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput UpperCamelCase__: List[str] = logging.get_logger(__name__) # pylint: disable=invalid-name def snake_case_ ( _lowerCAmelCase : Union[List, PIL.Image.Image, torch.Tensor] ) -> List[str]: warnings.warn( '''The preprocess method is deprecated and will be removed in a future version. Please''' ''' use VaeImageProcessor.preprocess instead''' , _lowerCAmelCase , ) if isinstance(_lowerCAmelCase , torch.Tensor ): return image elif isinstance(_lowerCAmelCase , PIL.Image.Image ): UpperCAmelCase : Optional[int] = [image] if isinstance(image[0] , PIL.Image.Image ): UpperCAmelCase , UpperCAmelCase : int = image[0].size UpperCAmelCase , UpperCAmelCase : Any = (x - x % 8 for x in (w, h)) # resize to integer multiple of 8 UpperCAmelCase : Optional[int] = [np.array(i.resize((w, h) , resample=PIL_INTERPOLATION['''lanczos'''] ) )[None, :] for i in image] UpperCAmelCase : Union[str, Any] = np.concatenate(_lowerCAmelCase , axis=0 ) UpperCAmelCase : str = np.array(_lowerCAmelCase ).astype(np.floataa ) / 2_5_5.0 UpperCAmelCase : Dict = image.transpose(0 , 3 , 1 , 2 ) UpperCAmelCase : Union[str, Any] = 2.0 * image - 1.0 UpperCAmelCase : Any = torch.from_numpy(_lowerCAmelCase ) elif isinstance(image[0] , torch.Tensor ): UpperCAmelCase : List[Any] = torch.cat(_lowerCAmelCase , dim=0 ) return image def snake_case_ ( _lowerCAmelCase : Union[List, PIL.Image.Image, torch.Tensor] ) -> Tuple: if isinstance(_lowerCAmelCase , torch.Tensor ): return mask elif isinstance(_lowerCAmelCase , PIL.Image.Image ): UpperCAmelCase : Optional[Any] = [mask] if isinstance(mask[0] , PIL.Image.Image ): UpperCAmelCase , UpperCAmelCase : Optional[Any] = mask[0].size UpperCAmelCase , UpperCAmelCase : str = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 UpperCAmelCase : Any = [np.array(m.convert('''L''' ).resize((w, h) , resample=PIL_INTERPOLATION['''nearest'''] ) )[None, :] for m in mask] UpperCAmelCase : Union[str, Any] = np.concatenate(_lowerCAmelCase , axis=0 ) UpperCAmelCase : List[Any] = mask.astype(np.floataa ) / 2_5_5.0 UpperCAmelCase : List[Any] = 0 UpperCAmelCase : Any = 1 UpperCAmelCase : List[Any] = torch.from_numpy(_lowerCAmelCase ) elif isinstance(mask[0] , torch.Tensor ): UpperCAmelCase : Dict = torch.cat(_lowerCAmelCase , dim=0 ) return mask class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = 42 lowerCamelCase__ = 42 def __init__( self : Tuple , __snake_case : Optional[Any] , __snake_case : str ) -> Union[str, Any]: super().__init__() self.register_modules(unet=__snake_case , scheduler=__snake_case ) @torch.no_grad() def __call__( self : Union[str, Any] , __snake_case : Union[torch.Tensor, PIL.Image.Image] , __snake_case : Union[torch.Tensor, PIL.Image.Image] , __snake_case : int = 250 , __snake_case : float = 0.0 , __snake_case : int = 10 , __snake_case : int = 10 , __snake_case : Optional[Union[torch.Generator, List[torch.Generator]]] = None , __snake_case : Optional[str] = "pil" , __snake_case : bool = True , ) -> Union[ImagePipelineOutput, Tuple]: UpperCAmelCase : int = image UpperCAmelCase : int = _preprocess_image(__snake_case ) UpperCAmelCase : int = original_image.to(device=self.device , dtype=self.unet.dtype ) UpperCAmelCase : Tuple = _preprocess_mask(__snake_case ) UpperCAmelCase : List[Any] = mask_image.to(device=self.device , dtype=self.unet.dtype ) UpperCAmelCase : Dict = original_image.shape[0] # sample gaussian noise to begin the loop if isinstance(__snake_case , __snake_case ) and len(__snake_case ) != batch_size: raise ValueError( F"""You have passed a list of generators of length {len(__snake_case )}, but requested an effective batch""" F""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) UpperCAmelCase : Union[str, Any] = original_image.shape UpperCAmelCase : Optional[Any] = randn_tensor(__snake_case , generator=__snake_case , device=self.device , dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(__snake_case , __snake_case , __snake_case , self.device ) UpperCAmelCase : Any = eta UpperCAmelCase : Optional[int] = self.scheduler.timesteps[0] + 1 UpperCAmelCase : Any = generator[0] if isinstance(__snake_case , __snake_case ) else generator for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): if t < t_last: # predict the noise residual UpperCAmelCase : Optional[int] = self.unet(__snake_case , __snake_case ).sample # compute previous image: x_t -> x_t-1 UpperCAmelCase : Union[str, Any] = self.scheduler.step(__snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case ).prev_sample else: # compute the reverse: x_t-1 -> x_t UpperCAmelCase : Tuple = self.scheduler.undo_step(__snake_case , __snake_case , __snake_case ) UpperCAmelCase : int = t UpperCAmelCase : Optional[int] = (image / 2 + 0.5).clamp(0 , 1 ) UpperCAmelCase : str = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": UpperCAmelCase : Union[str, Any] = self.numpy_to_pil(__snake_case ) if not return_dict: return (image,) return ImagePipelineOutput(images=__snake_case )
23
from __future__ import annotations def snake_case_ ( lowerCAmelCase_ : int , lowerCAmelCase_ : int ): if partitions <= 0: raise ValueError("""partitions must be a positive number!""" ) if partitions > number_of_bytes: raise ValueError("""partitions can not > number_of_bytes!""" ) __lowercase : Dict = number_of_bytes // partitions __lowercase : Union[str, Any] = [] for i in range(lowerCAmelCase_ ): __lowercase : str = i * bytes_per_partition + 1 __lowercase : List[Any] = ( number_of_bytes if i == partitions - 1 else (i + 1) * bytes_per_partition ) allocation_list.append(F"{start_bytes}-{end_bytes}" ) return allocation_list if __name__ == "__main__": import doctest doctest.testmod()
233
0
"""simple docstring""" from __future__ import annotations import unittest import numpy as np from transformers import LayoutLMConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.layoutlm.modeling_tf_layoutlm import ( TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMForMaskedLM, TFLayoutLMForQuestionAnswering, TFLayoutLMForSequenceClassification, TFLayoutLMForTokenClassification, TFLayoutLMModel, ) class _UpperCamelCase : '''simple docstring''' def __init__( self , __a , __a=13 , __a=7 , __a=True , __a=True , __a=True , __a=True , __a=99 , __a=32 , __a=2 , __a=4 , __a=37 , __a="gelu" , __a=0.1 , __a=0.1 , __a=5_12 , __a=16 , __a=2 , __a=0.0_2 , __a=3 , __a=4 , __a=None , __a=10_00 , ): __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = seq_length __lowerCAmelCase = is_training __lowerCAmelCase = use_input_mask __lowerCAmelCase = use_token_type_ids __lowerCAmelCase = use_labels __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = type_vocab_size __lowerCAmelCase = type_sequence_label_size __lowerCAmelCase = initializer_range __lowerCAmelCase = num_labels __lowerCAmelCase = num_choices __lowerCAmelCase = scope __lowerCAmelCase = range_bbox def snake_case ( self ): __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) # convert bbox to numpy since TF does not support item assignment __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ).numpy() # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: __lowerCAmelCase = bbox[i, j, 3] __lowerCAmelCase = bbox[i, j, 1] __lowerCAmelCase = t if bbox[i, j, 2] < bbox[i, j, 0]: __lowerCAmelCase = bbox[i, j, 2] __lowerCAmelCase = bbox[i, j, 0] __lowerCAmelCase = t __lowerCAmelCase = tf.convert_to_tensor(__a ) __lowerCAmelCase = None if self.use_input_mask: __lowerCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCAmelCase = None if self.use_token_type_ids: __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCAmelCase = None __lowerCAmelCase = None __lowerCAmelCase = None if self.use_labels: __lowerCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCAmelCase = ids_tensor([self.batch_size] , self.num_choices ) __lowerCAmelCase = LayoutLMConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def snake_case ( self , __a , __a , __a , __a , __a , __a , __a , __a ): __lowerCAmelCase = TFLayoutLMModel(config=__a ) __lowerCAmelCase = model(__a , __a , attention_mask=__a , token_type_ids=__a ) __lowerCAmelCase = model(__a , __a , token_type_ids=__a ) __lowerCAmelCase = model(__a , __a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def snake_case ( self , __a , __a , __a , __a , __a , __a , __a , __a ): __lowerCAmelCase = TFLayoutLMForMaskedLM(config=__a ) __lowerCAmelCase = model(__a , __a , attention_mask=__a , token_type_ids=__a , labels=__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case ( self , __a , __a , __a , __a , __a , __a , __a , __a ): __lowerCAmelCase = self.num_labels __lowerCAmelCase = TFLayoutLMForSequenceClassification(config=__a ) __lowerCAmelCase = model(__a , __a , attention_mask=__a , token_type_ids=__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case ( self , __a , __a , __a , __a , __a , __a , __a , __a ): __lowerCAmelCase = self.num_labels __lowerCAmelCase = TFLayoutLMForTokenClassification(config=__a ) __lowerCAmelCase = model(__a , __a , attention_mask=__a , token_type_ids=__a , labels=__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def snake_case ( self , __a , __a , __a , __a , __a , __a , __a , __a ): __lowerCAmelCase = TFLayoutLMForQuestionAnswering(config=__a ) __lowerCAmelCase = model(__a , __a , attention_mask=__a , token_type_ids=__a ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def snake_case ( self ): __lowerCAmelCase = self.prepare_config_and_inputs() ( ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ) = config_and_inputs __lowerCAmelCase = { "input_ids": input_ids, "bbox": bbox, "token_type_ids": token_type_ids, "attention_mask": input_mask, } return config, inputs_dict @require_tf class _UpperCamelCase ( lowerCAmelCase__ ,lowerCAmelCase__ ,unittest.TestCase ): '''simple docstring''' __UpperCAmelCase : List[Any] =( ( TFLayoutLMModel, TFLayoutLMForMaskedLM, TFLayoutLMForTokenClassification, TFLayoutLMForSequenceClassification, TFLayoutLMForQuestionAnswering, ) if is_tf_available() else () ) __UpperCAmelCase : List[Any] =( { """feature-extraction""": TFLayoutLMModel, """fill-mask""": TFLayoutLMForMaskedLM, """text-classification""": TFLayoutLMForSequenceClassification, """token-classification""": TFLayoutLMForTokenClassification, """zero-shot""": TFLayoutLMForSequenceClassification, } if is_tf_available() else {} ) __UpperCAmelCase : int =False __UpperCAmelCase : str =True __UpperCAmelCase : Optional[Any] =1_0 def snake_case ( self ): __lowerCAmelCase = TFLayoutLMModelTester(self ) __lowerCAmelCase = ConfigTester(self , config_class=__a , hidden_size=37 ) def snake_case ( self ): self.config_tester.run_common_tests() def snake_case ( self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a ) def snake_case ( self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__a ) def snake_case ( self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__a ) def snake_case ( self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__a ) def snake_case ( self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__a ) @slow def snake_case ( self ): for model_name in TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase = TFLayoutLMModel.from_pretrained(__a ) self.assertIsNotNone(__a ) @unittest.skip("Onnx compliancy broke with TF 2.10" ) def snake_case ( self ): pass def _lowerCamelCase ( ): '''simple docstring''' __lowerCAmelCase = tf.convert_to_tensor([[101,1019,1014,1016,1037,1_2849,4747,1004,1_4246,2278,5439,4524,5002,2930,2193,2930,4341,3208,1005,1055,2171,2848,1_1300,3531,102],[101,4070,4034,7020,1024,3058,1015,1013,2861,1013,6070,1_9274,2772,6205,2_7814,1_6147,1_6147,4343,2047,1_0283,1_0969,1_4389,1012,2338,102]] ) # noqa: E231 __lowerCAmelCase = tf.convert_to_tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],] ) # noqa: E231 __lowerCAmelCase = tf.convert_to_tensor([[[0,0,0,0],[423,237,440,251],[427,272,441,287],[419,115,437,129],[961,885,992,912],[256,38,330,58],[256,38,330,58],[336,42,353,57],[360,39,401,56],[360,39,401,56],[411,39,471,59],[479,41,528,59],[533,39,630,60],[67,113,134,131],[141,115,209,132],[68,149,133,166],[141,149,187,164],[195,148,287,165],[195,148,287,165],[195,148,287,165],[295,148,349,165],[441,149,492,166],[497,149,546,164],[64,201,125,218],[1000,1000,1000,1000]],[[0,0,0,0],[662,150,754,166],[665,199,742,211],[519,213,554,228],[519,213,554,228],[134,433,187,454],[130,467,204,480],[130,467,204,480],[130,467,204,480],[130,467,204,480],[130,467,204,480],[314,469,376,482],[504,684,582,706],[941,825,973,900],[941,825,973,900],[941,825,973,900],[941,825,973,900],[610,749,652,765],[130,659,168,672],[176,657,237,672],[238,657,312,672],[443,653,628,672],[443,653,628,672],[716,301,825,317],[1000,1000,1000,1000]]] ) # noqa: E231 __lowerCAmelCase = tf.convert_to_tensor([[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]] ) # noqa: E231 # these are sequence labels (i.e. at the token level) __lowerCAmelCase = tf.convert_to_tensor([[-100,10,10,10,9,1,-100,7,7,-100,7,7,4,2,5,2,8,8,-100,-100,5,0,3,2,-100],[-100,12,12,12,-100,12,10,-100,-100,-100,-100,10,12,9,-100,-100,-100,10,10,10,9,12,-100,10,-100]] ) # noqa: E231 # fmt: on return input_ids, attention_mask, bbox, token_type_ids, labels @require_tf class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def snake_case ( self ): __lowerCAmelCase = TFLayoutLMModel.from_pretrained("microsoft/layoutlm-base-uncased" ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = prepare_layoutlm_batch_inputs() # forward pass __lowerCAmelCase = model(input_ids=__a , bbox=__a , attention_mask=__a , token_type_ids=__a ) # test the sequence output on [0, :3, :3] __lowerCAmelCase = tf.convert_to_tensor( [[0.1_7_8_5, -0.1_9_4_7, -0.0_4_2_5], [-0.3_2_5_4, -0.2_8_0_7, 0.2_5_5_3], [-0.5_3_9_1, -0.3_3_2_2, 0.3_3_6_4]] , ) self.assertTrue(np.allclose(outputs.last_hidden_state[0, :3, :3] , __a , atol=1e-3 ) ) # test the pooled output on [1, :3] __lowerCAmelCase = tf.convert_to_tensor([-0.6_5_8_0, -0.0_2_1_4, 0.8_5_5_2] ) self.assertTrue(np.allclose(outputs.pooler_output[1, :3] , __a , atol=1e-3 ) ) @slow def snake_case ( self ): # initialize model with randomly initialized sequence classification head __lowerCAmelCase = TFLayoutLMForSequenceClassification.from_pretrained("microsoft/layoutlm-base-uncased" , num_labels=2 ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = prepare_layoutlm_batch_inputs() # forward pass __lowerCAmelCase = model( input_ids=__a , bbox=__a , attention_mask=__a , token_type_ids=__a , labels=tf.convert_to_tensor([1, 1] ) , ) # test whether we get a loss as a scalar __lowerCAmelCase = outputs.loss __lowerCAmelCase = (2,) self.assertEqual(loss.shape , __a ) # test the shape of the logits __lowerCAmelCase = outputs.logits __lowerCAmelCase = (2, 2) self.assertEqual(logits.shape , __a ) @slow def snake_case ( self ): # initialize model with randomly initialized token classification head __lowerCAmelCase = TFLayoutLMForTokenClassification.from_pretrained("microsoft/layoutlm-base-uncased" , num_labels=13 ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = prepare_layoutlm_batch_inputs() # forward pass __lowerCAmelCase = model( input_ids=__a , bbox=__a , attention_mask=__a , token_type_ids=__a , labels=__a ) # test the shape of the logits __lowerCAmelCase = outputs.logits __lowerCAmelCase = tf.convert_to_tensor((2, 25, 13) ) self.assertEqual(logits.shape , __a ) @slow def snake_case ( self ): # initialize model with randomly initialized token classification head __lowerCAmelCase = TFLayoutLMForQuestionAnswering.from_pretrained("microsoft/layoutlm-base-uncased" ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = prepare_layoutlm_batch_inputs() # forward pass __lowerCAmelCase = model(input_ids=__a , bbox=__a , attention_mask=__a , token_type_ids=__a ) # test the shape of the logits __lowerCAmelCase = tf.convert_to_tensor((2, 25) ) self.assertEqual(outputs.start_logits.shape , __a ) self.assertEqual(outputs.end_logits.shape , __a )
259
"""simple docstring""" import string def _lowerCamelCase ( _UpperCamelCase ): '''simple docstring''' __lowerCAmelCase = "" for i in sequence: __lowerCAmelCase = ord(_UpperCamelCase ) if 65 <= extract <= 90: output += chr(155 - extract ) elif 97 <= extract <= 122: output += chr(219 - extract ) else: output += i return output def _lowerCamelCase ( _UpperCamelCase ): '''simple docstring''' __lowerCAmelCase = string.ascii_letters __lowerCAmelCase = string.ascii_lowercase[::-1] + string.ascii_uppercase[::-1] return "".join( letters_reversed[letters.index(_UpperCamelCase )] if c in letters else c for c in sequence ) def _lowerCamelCase ( ): '''simple docstring''' from timeit import timeit print("Running performance benchmarks..." ) __lowerCAmelCase = "from string import printable ; from __main__ import atbash, atbash_slow" print(f"> atbash_slow(): {timeit('atbash_slow(printable)' , setup=_UpperCamelCase )} seconds" ) print(f"> atbash(): {timeit('atbash(printable)' , setup=_UpperCamelCase )} seconds" ) if __name__ == "__main__": for example in ("ABCDEFGH", "123GGjj", "testStringtest", "with space"): print(f'''{example} encrypted in atbash: {atbash(example)}''') benchmark()
259
1
import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.trocr.modeling_trocr import TrOCRDecoder, TrOCRForCausalLM @require_torch class UpperCAmelCase__ : '''simple docstring''' def __init__( self : Tuple , a_ : Optional[int] , a_ : Optional[int]=99 , a_ : Optional[int]=13 , a_ : List[Any]=16 , a_ : Optional[Any]=7 , a_ : Optional[int]=True , a_ : int=True , a_ : List[str]=True , a_ : List[Any]=False , a_ : int=True , a_ : List[str]=2 , a_ : str=32 , a_ : Any=4 , a_ : List[Any]=4 , a_ : List[str]=30 , a_ : Union[str, Any]=0 , a_ : int=1 , a_ : Tuple=2 , a_ : str=None , ): '''simple docstring''' __UpperCAmelCase : List[str] = parent __UpperCAmelCase : List[str] = batch_size __UpperCAmelCase : Tuple = decoder_seq_length # For common tests __UpperCAmelCase : str = self.decoder_seq_length __UpperCAmelCase : Union[str, Any] = is_training __UpperCAmelCase : Tuple = use_attention_mask __UpperCAmelCase : Any = use_labels __UpperCAmelCase : List[str] = vocab_size __UpperCAmelCase : Union[str, Any] = d_model __UpperCAmelCase : List[Any] = d_model __UpperCAmelCase : Tuple = decoder_layers __UpperCAmelCase : List[Any] = decoder_layers __UpperCAmelCase : Optional[Any] = decoder_ffn_dim __UpperCAmelCase : Any = decoder_attention_heads __UpperCAmelCase : str = decoder_attention_heads __UpperCAmelCase : Optional[Any] = eos_token_id __UpperCAmelCase : int = bos_token_id __UpperCAmelCase : Optional[int] = pad_token_id __UpperCAmelCase : List[Any] = decoder_start_token_id __UpperCAmelCase : Union[str, Any] = use_cache __UpperCAmelCase : Optional[int] = max_position_embeddings __UpperCAmelCase : Union[str, Any] = None __UpperCAmelCase : int = decoder_seq_length __UpperCAmelCase : List[str] = 2 __UpperCAmelCase : Union[str, Any] = 1 def snake_case__ ( self : Any ): '''simple docstring''' __UpperCAmelCase : Optional[Any] = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) __UpperCAmelCase : Optional[Any] = None if self.use_attention_mask: __UpperCAmelCase : Dict = ids_tensor([self.batch_size, self.decoder_seq_length] , vocab_size=2 ) __UpperCAmelCase : Dict = None if self.use_labels: __UpperCAmelCase : str = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) __UpperCAmelCase : Optional[int] = TrOCRConfig( vocab_size=self.vocab_size , d_model=self.d_model , decoder_layers=self.decoder_layers , decoder_ffn_dim=self.decoder_ffn_dim , decoder_attention_heads=self.decoder_attention_heads , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , use_cache=self.use_cache , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , max_position_embeddings=self.max_position_embeddings , ) return (config, input_ids, attention_mask, lm_labels) def snake_case__ ( self : Tuple , a_ : List[Any] , a_ : List[Any] , a_ : List[Any] , a_ : List[str] , ): '''simple docstring''' __UpperCAmelCase : Union[str, Any] = True __UpperCAmelCase : Dict = TrOCRDecoder(config=a_ ).to(a_ ).eval() __UpperCAmelCase : List[Any] = input_ids[:2] input_ids[input_ids == 0] += 1 # first forward pass __UpperCAmelCase : Optional[int] = model(a_ , use_cache=a_ ) __UpperCAmelCase : Union[str, Any] = model(a_ ) __UpperCAmelCase : Optional[Any] = model(a_ , use_cache=a_ ) self.parent.assertTrue(len(a_ ) == len(a_ ) ) self.parent.assertTrue(len(a_ ) == len(a_ ) + 1 ) __UpperCAmelCase : Any = outputs['''past_key_values'''] # create hypothetical next token and extent to next_input_ids __UpperCAmelCase : List[Any] = ids_tensor((2, 1) , config.vocab_size - 1 ) + 1 # append to next input_ids and __UpperCAmelCase : Optional[Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) __UpperCAmelCase : Union[str, Any] = model(a_ )['''last_hidden_state'''] __UpperCAmelCase : Dict = model(a_ , past_key_values=a_ )['''last_hidden_state'''] # select random slice __UpperCAmelCase : Any = ids_tensor((1,) , output_from_past.shape[-1] ).item() __UpperCAmelCase : str = output_from_no_past[:, next_input_ids.shape[-1] - 1, random_slice_idx].detach() __UpperCAmelCase : Optional[int] = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice assert torch.allclose(a_ , a_ , atol=1e-3 ) def snake_case__ ( self : Union[str, Any] ): '''simple docstring''' __UpperCAmelCase : str = self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : List[Any] = config_and_inputs __UpperCAmelCase : Tuple = {'''input_ids''': input_ids, '''attention_mask''': attention_mask} return config, inputs_dict @require_torch class UpperCAmelCase__ ( __UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,unittest.TestCase ): '''simple docstring''' UpperCamelCase = (TrOCRDecoder, TrOCRForCausalLM) if is_torch_available() else () UpperCamelCase = (TrOCRForCausalLM,) if is_torch_available() else () UpperCamelCase = {"""text-generation""": TrOCRForCausalLM} if is_torch_available() else {} UpperCamelCase = True UpperCamelCase = False def snake_case__ ( self : Any ): '''simple docstring''' __UpperCAmelCase : List[str] = TrOCRStandaloneDecoderModelTester(self , is_training=a_ ) __UpperCAmelCase : Dict = ConfigTester(self , config_class=a_ ) def snake_case__ ( self : Optional[int] ): '''simple docstring''' pass def snake_case__ ( self : Union[str, Any] ): '''simple docstring''' pass def snake_case__ ( self : Union[str, Any] ): '''simple docstring''' pass def snake_case__ ( self : Dict ): '''simple docstring''' self.config_tester.run_common_tests() def snake_case__ ( self : List[str] ): '''simple docstring''' __UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past(*a_ ) def snake_case__ ( self : List[Any] ): '''simple docstring''' return @unittest.skip('''The model doesn\'t support left padding''' ) # and it's not used enough to be worth fixing :) def snake_case__ ( self : str ): '''simple docstring''' pass
226
import numpy as np class UpperCAmelCase__ : '''simple docstring''' def __init__( self : List[Any] ): '''simple docstring''' __UpperCAmelCase : Optional[Any] = (0, 0) __UpperCAmelCase : List[str] = None __UpperCAmelCase : int = 0 __UpperCAmelCase : Union[str, Any] = 0 __UpperCAmelCase : List[Any] = 0 def __eq__( self : Optional[int] , a_ : List[Any] ): '''simple docstring''' return self.position == cell.position def snake_case__ ( self : List[Any] ): '''simple docstring''' print(self.position ) class UpperCAmelCase__ : '''simple docstring''' def __init__( self : str , a_ : Dict=(5, 5) ): '''simple docstring''' __UpperCAmelCase : str = np.zeros(a_ ) __UpperCAmelCase : List[Any] = world_size[0] __UpperCAmelCase : List[str] = world_size[1] def snake_case__ ( self : Tuple ): '''simple docstring''' print(self.w ) def snake_case__ ( self : Union[str, Any] , a_ : Tuple ): '''simple docstring''' __UpperCAmelCase : Optional[Any] = [ (-1, -1), (-1, 0), (-1, 1), (0, -1), (0, 1), (1, -1), (1, 0), (1, 1), ] __UpperCAmelCase : str = cell.position[0] __UpperCAmelCase : List[Any] = cell.position[1] __UpperCAmelCase : str = [] for n in neughbour_cord: __UpperCAmelCase : Optional[Any] = current_x + n[0] __UpperCAmelCase : Any = current_y + n[1] if 0 <= x < self.world_x_limit and 0 <= y < self.world_y_limit: __UpperCAmelCase : int = Cell() __UpperCAmelCase : List[str] = (x, y) __UpperCAmelCase : List[Any] = cell neighbours.append(a_ ) return neighbours def a ( _UpperCAmelCase : List[Any] , _UpperCAmelCase : Any , _UpperCAmelCase : int ): '''simple docstring''' __UpperCAmelCase : List[Any] = [] __UpperCAmelCase : Optional[Any] = [] _open.append(_UpperCAmelCase ) while _open: __UpperCAmelCase : str = np.argmin([n.f for n in _open] ) __UpperCAmelCase : Optional[Any] = _open[min_f] _closed.append(_open.pop(_UpperCAmelCase ) ) if current == goal: break for n in world.get_neigbours(_UpperCAmelCase ): for c in _closed: if c == n: continue __UpperCAmelCase : Optional[Any] = current.g + 1 __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = n.position __UpperCAmelCase , __UpperCAmelCase : Tuple = goal.position __UpperCAmelCase : Union[str, Any] = (ya - ya) ** 2 + (xa - xa) ** 2 __UpperCAmelCase : Optional[Any] = n.h + n.g for c in _open: if c == n and c.f < n.f: continue _open.append(_UpperCAmelCase ) __UpperCAmelCase : str = [] while current.parent is not None: path.append(current.position ) __UpperCAmelCase : int = current.parent path.append(current.position ) return path[::-1] if __name__ == "__main__": __A =Gridworld() # Start position and goal __A =Cell() __A =(0, 0) __A =Cell() __A =(4, 4) print(f'''path from {start.position} to {goal.position}''') __A =astar(world, start, goal) # Just for visual reasons. for i in s: __A =1 print(world.w)
226
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) _lowerCamelCase : Optional[Any] = '''▁''' _lowerCamelCase : Dict = {'''vocab_file''': '''spiece.model'''} _lowerCamelCase : Dict = { '''vocab_file''': { '''google/reformer-crime-and-punishment''': ( '''https://huggingface.co/google/reformer-crime-and-punishment/resolve/main/spiece.model''' ) } } _lowerCamelCase : Dict = { '''google/reformer-crime-and-punishment''': 524_288, } class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' _UpperCAmelCase : Optional[int] = VOCAB_FILES_NAMES _UpperCAmelCase : Any = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase : int = ["input_ids", "attention_mask"] def __init__( self : Optional[int] , lowercase : Tuple , lowercase : Tuple="</s>" , lowercase : Optional[Any]="<unk>" , lowercase : List[str]=[] , lowercase : Optional[Dict[str, Any]] = None , **lowercase : Optional[Any] , ): '''simple docstring''' _snake_case = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=lowercase , unk_token=lowercase , additional_special_tokens=lowercase , sp_model_kwargs=self.sp_model_kwargs , **lowercase , ) _snake_case = vocab_file _snake_case = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowercase ) @property def A ( self : int ): '''simple docstring''' return self.sp_model.get_piece_size() def A ( self : List[Any] ): '''simple docstring''' _snake_case = {self.convert_ids_to_tokens(lowercase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Dict ): '''simple docstring''' _snake_case = self.__dict__.copy() _snake_case = None return state def __setstate__( self : Dict , lowercase : List[str] ): '''simple docstring''' _snake_case = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): _snake_case = {} _snake_case = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def A ( self : List[str] , lowercase : str ): '''simple docstring''' return self.sp_model.encode(lowercase , out_type=lowercase ) def A ( self : Dict , lowercase : Optional[Any] ): '''simple docstring''' return self.sp_model.piece_to_id(lowercase ) def A ( self : Optional[int] , lowercase : List[str] ): '''simple docstring''' if index < self.sp_model.get_piece_size(): _snake_case = self.sp_model.IdToPiece(lowercase ) return token def A ( self : str , lowercase : Optional[Any] ): '''simple docstring''' _snake_case = [] _snake_case = '' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(lowercase ) + token _snake_case = [] else: current_sub_tokens.append(lowercase ) out_string += self.sp_model.decode(lowercase ) return out_string.strip() def A ( self : Tuple , lowercase : str , lowercase : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(lowercase ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return _snake_case = os.path.join( lowercase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowercase ) elif not os.path.isfile(self.vocab_file ): with open(lowercase , 'wb' ) as fi: _snake_case = self.sp_model.serialized_model_proto() fi.write(lowercase ) return (out_vocab_file,)
130
import collections.abc from typing import Optional, Tuple, Union import torch import torch.utils.checkpoint from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_poolformer import PoolFormerConfig _lowerCamelCase : List[Any] = logging.get_logger(__name__) # General docstring _lowerCamelCase : Dict = '''PoolFormerConfig''' # Base docstring _lowerCamelCase : int = '''sail/poolformer_s12''' _lowerCamelCase : Optional[Any] = [1, 512, 7, 7] # Image classification docstring _lowerCamelCase : Optional[int] = '''sail/poolformer_s12''' _lowerCamelCase : List[Any] = '''tabby, tabby cat''' _lowerCamelCase : List[str] = [ '''sail/poolformer_s12''', # See all PoolFormer models at https://huggingface.co/models?filter=poolformer ] def a_ ( __lowercase : List[Any] , __lowercase : float = 0.0 , __lowercase : bool = False ) -> Optional[int]: if drop_prob == 0.0 or not training: return input _snake_case = 1 - drop_prob _snake_case = (input.shape[0],) + (1,) * (input.ndim - 1) # work with diff dim tensors, not just 2D ConvNets _snake_case = keep_prob + torch.rand(__lowercase , dtype=input.dtype , device=input.device ) random_tensor.floor_() # binarize _snake_case = input.div(__lowercase ) * random_tensor return output class SCREAMING_SNAKE_CASE__ ( nn.Module ): '''simple docstring''' def __init__( self : Optional[Any] , lowercase : Optional[float] = None ): '''simple docstring''' super().__init__() _snake_case = drop_prob def A ( self : Any , lowercase : torch.Tensor ): '''simple docstring''' return drop_path(lowercase , self.drop_prob , self.training ) def A ( self : Tuple ): '''simple docstring''' return "p={}".format(self.drop_prob ) class SCREAMING_SNAKE_CASE__ ( nn.Module ): '''simple docstring''' def __init__( self : Tuple , lowercase : Dict , lowercase : Dict , lowercase : str , lowercase : int , lowercase : Optional[Any] , lowercase : str=None ): '''simple docstring''' super().__init__() _snake_case = patch_size if isinstance(lowercase , collections.abc.Iterable ) else (patch_size, patch_size) _snake_case = stride if isinstance(lowercase , collections.abc.Iterable ) else (stride, stride) _snake_case = padding if isinstance(lowercase , collections.abc.Iterable ) else (padding, padding) _snake_case = nn.Convad(lowercase , lowercase , kernel_size=lowercase , stride=lowercase , padding=lowercase ) _snake_case = norm_layer(lowercase ) if norm_layer else nn.Identity() def A ( self : int , lowercase : Union[str, Any] ): '''simple docstring''' _snake_case = self.projection(lowercase ) _snake_case = self.norm(lowercase ) return embeddings class SCREAMING_SNAKE_CASE__ ( nn.GroupNorm ): '''simple docstring''' def __init__( self : Dict , lowercase : List[Any] , **lowercase : str ): '''simple docstring''' super().__init__(1 , lowercase , **lowercase ) class SCREAMING_SNAKE_CASE__ ( nn.Module ): '''simple docstring''' def __init__( self : int , lowercase : List[Any] ): '''simple docstring''' super().__init__() _snake_case = nn.AvgPoolad(lowercase , stride=1 , padding=pool_size // 2 , count_include_pad=lowercase ) def A ( self : int , lowercase : List[str] ): '''simple docstring''' return self.pool(lowercase ) - hidden_states class SCREAMING_SNAKE_CASE__ ( nn.Module ): '''simple docstring''' def __init__( self : Dict , lowercase : Tuple , lowercase : str , lowercase : Optional[Any] , lowercase : Union[str, Any] ): '''simple docstring''' super().__init__() _snake_case = nn.Convad(lowercase , lowercase , 1 ) _snake_case = nn.Convad(lowercase , lowercase , 1 ) _snake_case = PoolFormerDropPath(lowercase ) if isinstance(config.hidden_act , lowercase ): _snake_case = ACTaFN[config.hidden_act] else: _snake_case = config.hidden_act def A ( self : Optional[int] , lowercase : str ): '''simple docstring''' _snake_case = self.conva(lowercase ) _snake_case = self.act_fn(lowercase ) _snake_case = self.drop(lowercase ) _snake_case = self.conva(lowercase ) _snake_case = self.drop(lowercase ) return hidden_states class SCREAMING_SNAKE_CASE__ ( nn.Module ): '''simple docstring''' def __init__( self : str , lowercase : Tuple , lowercase : int , lowercase : str , lowercase : Union[str, Any] , lowercase : str , lowercase : Dict ): '''simple docstring''' super().__init__() _snake_case = PoolFormerPooling(lowercase ) _snake_case = PoolFormerOutput(lowercase , lowercase , lowercase , lowercase ) _snake_case = PoolFormerGroupNorm(lowercase ) _snake_case = PoolFormerGroupNorm(lowercase ) # Useful for training neural nets _snake_case = PoolFormerDropPath(lowercase ) if drop_path > 0.0 else nn.Identity() _snake_case = config.use_layer_scale if config.use_layer_scale: _snake_case = nn.Parameter( config.layer_scale_init_value * torch.ones((lowercase) ) , requires_grad=lowercase ) _snake_case = nn.Parameter( config.layer_scale_init_value * torch.ones((lowercase) ) , requires_grad=lowercase ) def A ( self : Optional[int] , lowercase : Union[str, Any] ): '''simple docstring''' if self.use_layer_scale: _snake_case = self.pooling(self.before_norm(lowercase ) ) _snake_case = self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * pooling_output # First residual connection _snake_case = hidden_states + self.drop_path(lowercase ) _snake_case = () _snake_case = self.output(self.after_norm(lowercase ) ) _snake_case = self.layer_scale_a.unsqueeze(-1 ).unsqueeze(-1 ) * layer_output # Second residual connection _snake_case = hidden_states + self.drop_path(lowercase ) _snake_case = (output,) + outputs return outputs else: _snake_case = self.drop_path(self.pooling(self.before_norm(lowercase ) ) ) # First residual connection _snake_case = pooling_output + hidden_states _snake_case = () # Second residual connection inside the PoolFormerOutput block _snake_case = self.drop_path(self.output(self.after_norm(lowercase ) ) ) _snake_case = hidden_states + layer_output _snake_case = (output,) + outputs return outputs class SCREAMING_SNAKE_CASE__ ( nn.Module ): '''simple docstring''' def __init__( self : Optional[Any] , lowercase : Optional[int] ): '''simple docstring''' super().__init__() _snake_case = config # stochastic depth decay rule _snake_case = [x.item() for x in torch.linspace(0 , config.drop_path_rate , sum(config.depths ) )] # patch embeddings _snake_case = [] for i in range(config.num_encoder_blocks ): embeddings.append( PoolFormerEmbeddings( patch_size=config.patch_sizes[i] , stride=config.strides[i] , padding=config.padding[i] , num_channels=config.num_channels if i == 0 else config.hidden_sizes[i - 1] , hidden_size=config.hidden_sizes[i] , ) ) _snake_case = nn.ModuleList(lowercase ) # Transformer blocks _snake_case = [] _snake_case = 0 for i in range(config.num_encoder_blocks ): # each block consists of layers _snake_case = [] if i != 0: cur += config.depths[i - 1] for j in range(config.depths[i] ): layers.append( PoolFormerLayer( lowercase , num_channels=config.hidden_sizes[i] , pool_size=config.pool_size , hidden_size=config.hidden_sizes[i] , intermediate_size=int(config.hidden_sizes[i] * config.mlp_ratio ) , drop_path=dpr[cur + j] , ) ) blocks.append(nn.ModuleList(lowercase ) ) _snake_case = nn.ModuleList(lowercase ) def A ( self : Any , lowercase : List[str] , lowercase : str=False , lowercase : Tuple=True ): '''simple docstring''' _snake_case = () if output_hidden_states else None _snake_case = pixel_values for idx, layers in enumerate(zip(self.patch_embeddings , self.block ) ): _snake_case , _snake_case = layers # Get patch embeddings from hidden_states _snake_case = embedding_layer(lowercase ) # Send the embeddings through the blocks for _, blk in enumerate(lowercase ): _snake_case = blk(lowercase ) _snake_case = layer_outputs[0] if output_hidden_states: _snake_case = all_hidden_states + (hidden_states,) if not return_dict: return tuple(v for v in [hidden_states, all_hidden_states] if v is not None ) return BaseModelOutputWithNoAttention(last_hidden_state=lowercase , hidden_states=lowercase ) class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' _UpperCAmelCase : List[Any] = PoolFormerConfig _UpperCAmelCase : Optional[int] = "poolformer" _UpperCAmelCase : str = "pixel_values" _UpperCAmelCase : int = True def A ( self : Tuple , lowercase : str ): '''simple docstring''' if isinstance(lowercase , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(lowercase , nn.LayerNorm ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) def A ( self : Optional[Any] , lowercase : str , lowercase : Dict=False ): '''simple docstring''' if isinstance(lowercase , lowercase ): _snake_case = value _lowerCamelCase : Optional[Any] = r''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`PoolFormerConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' _lowerCamelCase : Tuple = r''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`PoolFormerImageProcessor.__call__`] for details. ''' @add_start_docstrings( "The bare PoolFormer Model transformer outputting raw hidden-states without any specific head on top." ,UpperCAmelCase ,) class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' def __init__( self : str , lowercase : List[Any] ): '''simple docstring''' super().__init__(lowercase ) _snake_case = config _snake_case = PoolFormerEncoder(lowercase ) # Initialize weights and apply final processing self.post_init() def A ( self : List[str] ): '''simple docstring''' return self.embeddings.patch_embeddings @add_start_docstrings_to_model_forward(lowercase ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=lowercase , config_class=_CONFIG_FOR_DOC , modality='vision' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def A ( self : Tuple , lowercase : Optional[torch.FloatTensor] = None , lowercase : Optional[bool] = None , lowercase : Optional[bool] = None , ): '''simple docstring''' _snake_case = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _snake_case = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError('You have to specify pixel_values' ) _snake_case = self.encoder( lowercase , output_hidden_states=lowercase , return_dict=lowercase , ) _snake_case = encoder_outputs[0] if not return_dict: return (sequence_output, None) + encoder_outputs[1:] return BaseModelOutputWithNoAttention( last_hidden_state=lowercase , hidden_states=encoder_outputs.hidden_states , ) class SCREAMING_SNAKE_CASE__ ( nn.Module ): '''simple docstring''' def __init__( self : List[str] , lowercase : Union[str, Any] ): '''simple docstring''' super().__init__() _snake_case = nn.Linear(config.hidden_size , config.hidden_size ) def A ( self : Optional[Any] , lowercase : Optional[int] ): '''simple docstring''' _snake_case = self.dense(lowercase ) return output @add_start_docstrings( "\n PoolFormer Model transformer with an image classification head on top\n " ,UpperCAmelCase ,) class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' def __init__( self : str , lowercase : Any ): '''simple docstring''' super().__init__(lowercase ) _snake_case = config.num_labels _snake_case = PoolFormerModel(lowercase ) # Final norm _snake_case = PoolFormerGroupNorm(config.hidden_sizes[-1] ) # Classifier head _snake_case = ( nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(lowercase ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=lowercase , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def A ( self : Union[str, Any] , lowercase : Optional[torch.FloatTensor] = None , lowercase : Optional[torch.LongTensor] = None , lowercase : Optional[bool] = None , lowercase : Optional[bool] = None , ): '''simple docstring''' _snake_case = return_dict if return_dict is not None else self.config.use_return_dict _snake_case = self.poolformer( lowercase , output_hidden_states=lowercase , return_dict=lowercase , ) _snake_case = outputs[0] _snake_case = self.classifier(self.norm(lowercase ).mean([-2, -1] ) ) _snake_case = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: _snake_case = 'regression' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): _snake_case = 'single_label_classification' else: _snake_case = 'multi_label_classification' if self.config.problem_type == "regression": _snake_case = MSELoss() if self.num_labels == 1: _snake_case = loss_fct(logits.squeeze() , labels.squeeze() ) else: _snake_case = loss_fct(lowercase , lowercase ) elif self.config.problem_type == "single_label_classification": _snake_case = CrossEntropyLoss() _snake_case = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": _snake_case = BCEWithLogitsLoss() _snake_case = loss_fct(lowercase , lowercase ) if not return_dict: _snake_case = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=lowercase , logits=lowercase , hidden_states=outputs.hidden_states )
130
1
"""simple docstring""" from argparse import ArgumentParser from . import BaseTransformersCLICommand def __UpperCAmelCase ( __UpperCamelCase ): return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class UpperCAmelCase_ ( snake_case ): @staticmethod def _lowerCamelCase ( UpperCamelCase_ ) -> str: __lowercase : Tuple = parser.add_parser('''download''' ) download_parser.add_argument( '''--cache-dir''' , type=UpperCamelCase_ , default=UpperCamelCase_ , help='''Path to location to store the models''' ) download_parser.add_argument( '''--force''' , action='''store_true''' , help='''Force the model to be download even if already in cache-dir''' ) download_parser.add_argument( '''--trust-remote-code''' , action='''store_true''' , help='''Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you\'ve reviewed the code as it will execute on your local machine''' , ) download_parser.add_argument('''model''' , type=UpperCamelCase_ , help='''Name of the model to download''' ) download_parser.set_defaults(func=UpperCamelCase_ ) def __init__( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> str: __lowercase : Dict = model __lowercase : Union[str, Any] = cache __lowercase : List[str] = force __lowercase : int = trust_remote_code def _lowerCamelCase ( self ) -> Optional[Any]: from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code )
249
"""simple docstring""" import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import SegformerImageProcessor, SwinConfig, UperNetConfig, UperNetForSemanticSegmentation def __UpperCAmelCase ( __UpperCamelCase ): __lowercase : Optional[int] = 3_84 __lowercase : str = 7 if "tiny" in model_name: __lowercase : List[str] = 96 __lowercase : Any = (2, 2, 6, 2) __lowercase : Dict = (3, 6, 12, 24) elif "small" in model_name: __lowercase : str = 96 __lowercase : Optional[int] = (2, 2, 18, 2) __lowercase : Tuple = (3, 6, 12, 24) elif "base" in model_name: __lowercase : Tuple = 1_28 __lowercase : Tuple = (2, 2, 18, 2) __lowercase : int = (4, 8, 16, 32) __lowercase : str = 12 __lowercase : Any = 5_12 elif "large" in model_name: __lowercase : List[str] = 1_92 __lowercase : List[Any] = (2, 2, 18, 2) __lowercase : Optional[Any] = (6, 12, 24, 48) __lowercase : Optional[int] = 12 __lowercase : Optional[Any] = 7_68 # set label information __lowercase : Any = 1_50 __lowercase : Tuple = '''huggingface/label-files''' __lowercase : int = '''ade20k-id2label.json''' __lowercase : Union[str, Any] = json.load(open(hf_hub_download(__UpperCamelCase , __UpperCamelCase , repo_type='''dataset''' ) , '''r''' ) ) __lowercase : Union[str, Any] = {int(__UpperCamelCase ): v for k, v in idalabel.items()} __lowercase : Optional[Any] = {v: k for k, v in idalabel.items()} __lowercase : Any = SwinConfig( embed_dim=__UpperCamelCase , depths=__UpperCamelCase , num_heads=__UpperCamelCase , window_size=__UpperCamelCase , out_features=['''stage1''', '''stage2''', '''stage3''', '''stage4'''] , ) __lowercase : List[Any] = UperNetConfig( backbone_config=__UpperCamelCase , auxiliary_in_channels=__UpperCamelCase , num_labels=__UpperCamelCase , idalabel=__UpperCamelCase , labelaid=__UpperCamelCase , ) return config def __UpperCAmelCase ( __UpperCamelCase ): __lowercase : str = [] # fmt: off # stem rename_keys.append(('''backbone.patch_embed.projection.weight''', '''backbone.embeddings.patch_embeddings.projection.weight''') ) rename_keys.append(('''backbone.patch_embed.projection.bias''', '''backbone.embeddings.patch_embeddings.projection.bias''') ) rename_keys.append(('''backbone.patch_embed.norm.weight''', '''backbone.embeddings.norm.weight''') ) rename_keys.append(('''backbone.patch_embed.norm.bias''', '''backbone.embeddings.norm.bias''') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((f"""backbone.stages.{i}.blocks.{j}.norm1.weight""", f"""backbone.encoder.layers.{i}.blocks.{j}.layernorm_before.weight""") ) rename_keys.append((f"""backbone.stages.{i}.blocks.{j}.norm1.bias""", f"""backbone.encoder.layers.{i}.blocks.{j}.layernorm_before.bias""") ) rename_keys.append((f"""backbone.stages.{i}.blocks.{j}.attn.w_msa.relative_position_bias_table""", f"""backbone.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table""") ) rename_keys.append((f"""backbone.stages.{i}.blocks.{j}.attn.w_msa.relative_position_index""", f"""backbone.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index""") ) rename_keys.append((f"""backbone.stages.{i}.blocks.{j}.attn.w_msa.proj.weight""", f"""backbone.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight""") ) rename_keys.append((f"""backbone.stages.{i}.blocks.{j}.attn.w_msa.proj.bias""", f"""backbone.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias""") ) rename_keys.append((f"""backbone.stages.{i}.blocks.{j}.norm2.weight""", f"""backbone.encoder.layers.{i}.blocks.{j}.layernorm_after.weight""") ) rename_keys.append((f"""backbone.stages.{i}.blocks.{j}.norm2.bias""", f"""backbone.encoder.layers.{i}.blocks.{j}.layernorm_after.bias""") ) rename_keys.append((f"""backbone.stages.{i}.blocks.{j}.ffn.layers.0.0.weight""", f"""backbone.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight""") ) rename_keys.append((f"""backbone.stages.{i}.blocks.{j}.ffn.layers.0.0.bias""", f"""backbone.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias""") ) rename_keys.append((f"""backbone.stages.{i}.blocks.{j}.ffn.layers.1.weight""", f"""backbone.encoder.layers.{i}.blocks.{j}.output.dense.weight""") ) rename_keys.append((f"""backbone.stages.{i}.blocks.{j}.ffn.layers.1.bias""", f"""backbone.encoder.layers.{i}.blocks.{j}.output.dense.bias""") ) if i < 3: rename_keys.append((f"""backbone.stages.{i}.downsample.reduction.weight""", f"""backbone.encoder.layers.{i}.downsample.reduction.weight""") ) rename_keys.append((f"""backbone.stages.{i}.downsample.norm.weight""", f"""backbone.encoder.layers.{i}.downsample.norm.weight""") ) rename_keys.append((f"""backbone.stages.{i}.downsample.norm.bias""", f"""backbone.encoder.layers.{i}.downsample.norm.bias""") ) rename_keys.append((f"""backbone.norm{i}.weight""", f"""backbone.hidden_states_norms.stage{i+1}.weight""") ) rename_keys.append((f"""backbone.norm{i}.bias""", f"""backbone.hidden_states_norms.stage{i+1}.bias""") ) # decode head rename_keys.extend( [ ('''decode_head.conv_seg.weight''', '''decode_head.classifier.weight'''), ('''decode_head.conv_seg.bias''', '''decode_head.classifier.bias'''), ('''auxiliary_head.conv_seg.weight''', '''auxiliary_head.classifier.weight'''), ('''auxiliary_head.conv_seg.bias''', '''auxiliary_head.classifier.bias'''), ] ) # fmt: on return rename_keys def __UpperCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): __lowercase : Any = dct.pop(__UpperCamelCase ) __lowercase : Any = val def __UpperCAmelCase ( __UpperCamelCase , __UpperCamelCase ): __lowercase : Optional[Any] = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): __lowercase : Optional[Any] = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) __lowercase : Dict = state_dict.pop(f"""backbone.stages.{i}.blocks.{j}.attn.w_msa.qkv.weight""" ) __lowercase : int = state_dict.pop(f"""backbone.stages.{i}.blocks.{j}.attn.w_msa.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict __lowercase : List[Any] = in_proj_weight[:dim, :] __lowercase : Tuple = in_proj_bias[: dim] __lowercase : List[Any] = in_proj_weight[ dim : dim * 2, : ] __lowercase : int = in_proj_bias[ dim : dim * 2 ] __lowercase : str = in_proj_weight[ -dim :, : ] __lowercase : List[Any] = in_proj_bias[-dim :] # fmt: on def __UpperCAmelCase ( __UpperCamelCase ): __lowercase ,__lowercase : str = x.shape __lowercase : List[str] = x.reshape(__UpperCamelCase , 4 , in_channel // 4 ) __lowercase : Dict = x[:, [0, 2, 1, 3], :].transpose(1 , 2 ).reshape(__UpperCamelCase , __UpperCamelCase ) return x def __UpperCAmelCase ( __UpperCamelCase ): __lowercase ,__lowercase : Optional[int] = x.shape __lowercase : Union[str, Any] = x.reshape(__UpperCamelCase , in_channel // 4 , 4 ) __lowercase : int = x[:, :, [0, 2, 1, 3]].transpose(1 , 2 ).reshape(__UpperCamelCase , __UpperCamelCase ) return x def __UpperCAmelCase ( __UpperCamelCase ): __lowercase : int = x.shape[0] __lowercase : List[str] = x.reshape(4 , in_channel // 4 ) __lowercase : Any = x[[0, 2, 1, 3], :].transpose(0 , 1 ).reshape(__UpperCamelCase ) return x def __UpperCAmelCase ( __UpperCamelCase ): __lowercase : Union[str, Any] = x.shape[0] __lowercase : List[str] = x.reshape(in_channel // 4 , 4 ) __lowercase : Union[str, Any] = x[:, [0, 2, 1, 3]].transpose(0 , 1 ).reshape(__UpperCamelCase ) return x def __UpperCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): __lowercase : List[Any] = { '''upernet-swin-tiny''': '''https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_tiny_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K/upernet_swin_tiny_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K_20210531_112542-e380ad3e.pth''', '''upernet-swin-small''': '''https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_small_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K/upernet_swin_small_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K_20210526_192015-ee2fff1c.pth''', '''upernet-swin-base''': '''https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_base_patch4_window12_512x512_160k_ade20k_pretrain_384x384_22K/upernet_swin_base_patch4_window12_512x512_160k_ade20k_pretrain_384x384_22K_20210531_125459-429057bf.pth''', '''upernet-swin-large''': '''https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_large_patch4_window12_512x512_pretrain_384x384_22K_160k_ade20k/upernet_swin_large_patch4_window12_512x512_pretrain_384x384_22K_160k_ade20k_20220318_091743-9ba68901.pth''', } __lowercase : Any = model_name_to_url[model_name] __lowercase : Any = torch.hub.load_state_dict_from_url(__UpperCamelCase , map_location='''cpu''' , file_name=__UpperCamelCase )[ '''state_dict''' ] for name, param in state_dict.items(): print(__UpperCamelCase , param.shape ) __lowercase : Tuple = get_upernet_config(__UpperCamelCase ) __lowercase : List[Any] = UperNetForSemanticSegmentation(__UpperCamelCase ) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): __lowercase : Optional[Any] = state_dict.pop(__UpperCamelCase ) if "bn" in key: __lowercase : List[Any] = key.replace('''bn''' , '''batch_norm''' ) __lowercase : Optional[Any] = val # rename keys __lowercase : Tuple = create_rename_keys(__UpperCamelCase ) for src, dest in rename_keys: rename_key(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) read_in_q_k_v(__UpperCamelCase , config.backbone_config ) # fix downsample parameters for key, value in state_dict.items(): if "downsample" in key: if "reduction" in key: __lowercase : Optional[Any] = reverse_correct_unfold_reduction_order(__UpperCamelCase ) if "norm" in key: __lowercase : Optional[Any] = reverse_correct_unfold_norm_order(__UpperCamelCase ) model.load_state_dict(__UpperCamelCase ) # verify on image __lowercase : Any = '''https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg''' __lowercase : str = Image.open(requests.get(__UpperCamelCase , stream=__UpperCamelCase ).raw ).convert('''RGB''' ) __lowercase : Union[str, Any] = SegformerImageProcessor() __lowercase : int = processor(__UpperCamelCase , return_tensors='''pt''' ).pixel_values with torch.no_grad(): __lowercase : List[Any] = model(__UpperCamelCase ) __lowercase : Union[str, Any] = outputs.logits print(logits.shape ) print('''First values of logits:''' , logits[0, 0, :3, :3] ) # assert values if model_name == "upernet-swin-tiny": __lowercase : Tuple = torch.tensor( [[-7.5_958, -7.5_958, -7.4_302], [-7.5_958, -7.5_958, -7.4_302], [-7.4_797, -7.4_797, -7.3_068]] ) elif model_name == "upernet-swin-small": __lowercase : Optional[Any] = torch.tensor( [[-7.1_921, -7.1_921, -6.9_532], [-7.1_921, -7.1_921, -6.9_532], [-7.0_908, -7.0_908, -6.8_534]] ) elif model_name == "upernet-swin-base": __lowercase : Optional[int] = torch.tensor( [[-6.5_851, -6.5_851, -6.4_330], [-6.5_851, -6.5_851, -6.4_330], [-6.4_763, -6.4_763, -6.3_254]] ) elif model_name == "upernet-swin-large": __lowercase : Any = torch.tensor( [[-7.5_297, -7.5_297, -7.3_802], [-7.5_297, -7.5_297, -7.3_802], [-7.4_044, -7.4_044, -7.2_586]] ) print('''Logits:''' , outputs.logits[0, 0, :3, :3] ) assert torch.allclose(outputs.logits[0, 0, :3, :3] , __UpperCamelCase , atol=1e-4 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(f"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__UpperCamelCase ) print(f"""Saving processor to {pytorch_dump_folder_path}""" ) processor.save_pretrained(__UpperCamelCase ) if push_to_hub: print(f"""Pushing model and processor for {model_name} to hub""" ) model.push_to_hub(f"""openmmlab/{model_name}""" ) processor.push_to_hub(f"""openmmlab/{model_name}""" ) if __name__ == "__main__": a_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='upernet-swin-tiny', type=str, choices=[F"upernet-swin-{size}" for size in ['tiny', 'small', 'base', 'large']], help='Name of the Swin + UperNet model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) a_ = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
249
1
import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeqaSeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging a__ : List[Any] = logging.get_logger(__name__) def UpperCAmelCase_( a__ , a__ , a__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = nn.ModuleList([src_layers[i] for i in layers_to_copy] ) assert len(a__ ) == len(a__ ), F"""{len(a__ )} != {len(a__ )}""" dest_layers.load_state_dict(layers_to_copy.state_dict() ) a__ : List[str] = { # maps num layers in teacher -> num_layers in student -> which teacher layers to copy. # 12: bart, 16: pegasus, 6: marian/Helsinki-NLP 12: { 1: [0], # This says that if the teacher has 12 layers and the student has 1, copy layer 0 of the teacher 2: [0, 6], 3: [0, 6, 11], 4: [0, 4, 8, 11], 6: [0, 2, 4, 7, 9, 11], 9: [0, 1, 2, 4, 5, 7, 9, 10, 11], 12: list(range(12)), }, 16: { # maps num layers in student -> which teacher layers to copy 1: [0], 2: [0, 15], 3: [0, 8, 15], 4: [0, 5, 10, 15], 6: [0, 3, 6, 9, 12, 15], 8: [0, 2, 4, 6, 8, 10, 12, 15], 9: [0, 1, 3, 5, 7, 9, 11, 13, 15], 12: [0, 1, 2, 3, 4, 5, 6, 7, 9, 11, 13, 15], 16: list(range(16)), }, 6: {1: [0], 2: [0, 5], 3: [0, 2, 5], 4: [0, 1, 3, 5], 6: list(range(6))}, } a__ : List[str] = { # maps num layers in student -> which teacher layers to copy. 6: {1: [5], 2: [3, 5], 3: [1, 4, 5], 4: [1, 2, 4, 5]}, 12: {1: [11], 2: [5, 11], 3: [3, 7, 11], 6: [1, 3, 5, 8, 10, 11]}, 16: {1: [15], 4: [4, 9, 12, 15], 8: [1, 3, 5, 7, 9, 11, 13, 15]}, } def UpperCAmelCase_( a__ , a__ ): """simple docstring""" try: SCREAMING_SNAKE_CASE : Tuple = LAYERS_TO_COPY[n_teacher][n_student] return val except KeyError: if n_student != n_teacher: warnings.warn( F"""no hardcoded layers to copy for teacher {n_teacher} -> student {n_student}, defaulting to first""" F""" {n_student}""" ) return list(range(a__ ) ) def UpperCAmelCase_( a__ , a__ ): """simple docstring""" if n_student > n_teacher: raise ValueError(F"""Cannot perform intermediate supervision for student {n_student} > teacher {n_teacher}""" ) elif n_teacher == n_student: return list(range(a__ ) ) elif n_student == 1: return [n_teacher - 1] else: return LAYERS_TO_SUPERVISE[n_teacher][n_student] def UpperCAmelCase_( a__ , a__ = "student" , a__ = None , a__ = None , a__=False , a__=None , a__=None , **a__ , ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = '''encoder_layers and decoder_layers cannot be both None-- you would just have an identical teacher.''' assert (e is not None) or (d is not None), _msg if isinstance(a__ , a__ ): AutoTokenizer.from_pretrained(a__ ).save_pretrained(a__ ) # purely for convenience SCREAMING_SNAKE_CASE : Dict = AutoModelForSeqaSeqLM.from_pretrained(a__ ).eval() else: assert isinstance(a__ , a__ ), F"""teacher must be a model or string got type {type(a__ )}""" SCREAMING_SNAKE_CASE : List[Any] = teacher.config.to_diff_dict() try: SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Optional[Any] = teacher.config.encoder_layers, teacher.config.decoder_layers if e is None: SCREAMING_SNAKE_CASE : Optional[Any] = teacher_e if d is None: SCREAMING_SNAKE_CASE : Any = teacher_d init_kwargs.update({'''encoder_layers''': e, '''decoder_layers''': d} ) except AttributeError: # T5 if hasattr(teacher.config , '''num_encoder_layers''' ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Optional[int] = teacher.config.num_encoder_layers, teacher.config.num_decoder_layers else: SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Tuple = teacher.config.num_layers, teacher.config.num_decoder_layers if e is None: SCREAMING_SNAKE_CASE : List[Any] = teacher_e if d is None: SCREAMING_SNAKE_CASE : List[str] = teacher_d if hasattr(teacher.config , '''num_encoder_layers''' ): init_kwargs.update({'''num_encoder_layers''': e, '''num_decoder_layers''': d} ) else: init_kwargs.update({'''num_layers''': e, '''num_decoder_layers''': d} ) # Kwargs to instantiate student: teacher kwargs with updated layer numbers + **extra_config_kwargs init_kwargs.update(a__ ) # Copy weights SCREAMING_SNAKE_CASE : Dict = teacher.config_class(**a__ ) SCREAMING_SNAKE_CASE : int = AutoModelForSeqaSeqLM.from_config(a__ ) # Start by copying the full teacher state dict this will copy the first N teacher layers to the student. SCREAMING_SNAKE_CASE : Union[str, Any] = student.load_state_dict(teacher.state_dict() , strict=a__ ) assert info.missing_keys == [], info.missing_keys # every student key should have a teacher keys. if copy_first_teacher_layers: # Our copying is done. We just log and save SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : List[str] = list(range(a__ ) ), list(range(a__ ) ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to""" F""" {save_path}""" ) student.save_pretrained(a__ ) return student, e_layers_to_copy, d_layers_to_copy # Decide which layers of the teacher to copy. Not exactly alternating -- we try to keep first and last layer. if e_layers_to_copy is None: SCREAMING_SNAKE_CASE : List[int] = pick_layers_to_copy(a__ , a__ ) if d_layers_to_copy is None: SCREAMING_SNAKE_CASE : List[int] = pick_layers_to_copy(a__ , a__ ) try: if hasattr( a__ , '''prophetnet''' ): # For ProphetNet, student.model.encoder.layers is called student.prophetnet.encoder.layers copy_layers(teacher.prophetnet.encoder.layers , student.prophetnet.encoder.layers , a__ ) copy_layers(teacher.prophetnet.decoder.layers , student.prophetnet.decoder.layers , a__ ) else: copy_layers(teacher.model.encoder.layers , student.model.encoder.layers , a__ ) copy_layers(teacher.model.decoder.layers , student.model.decoder.layers , a__ ) except AttributeError: # For t5, student.model.encoder.layers is called student.encoder.block copy_layers(teacher.encoder.block , student.encoder.block , a__ ) copy_layers(teacher.decoder.block , student.decoder.block , a__ ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to {save_path}""" ) SCREAMING_SNAKE_CASE : Tuple = { '''teacher_type''': teacher.config.model_type, '''copied_encoder_layers''': e_layers_to_copy, '''copied_decoder_layers''': d_layers_to_copy, } student.save_pretrained(a__ ) # Save information about copying for easier reproducibility return student, e_layers_to_copy, d_layers_to_copy if __name__ == "__main__": fire.Fire(create_student_by_copying_alternating_layers)
19
import datasets from .evaluate import evaluate a__ : Dict = '''\ @article{hendrycks2021cuad, title={CUAD: An Expert-Annotated NLP Dataset for Legal Contract Review}, author={Dan Hendrycks and Collin Burns and Anya Chen and Spencer Ball}, journal={arXiv preprint arXiv:2103.06268}, year={2021} } ''' a__ : List[str] = ''' This metric wrap the official scoring script for version 1 of the Contract Understanding Atticus Dataset (CUAD). Contract Understanding Atticus Dataset (CUAD) v1 is a corpus of more than 13,000 labels in 510 commercial legal contracts that have been manually labeled to identify 41 categories of important clauses that lawyers look for when reviewing contracts in connection with corporate transactions. ''' a__ : List[Any] = ''' Computes CUAD scores (EM, F1, AUPR, Precision@80%Recall, and Precision@90%Recall). Args: predictions: List of question-answers dictionaries with the following key-values: - \'id\': id of the question-answer pair as given in the references (see below) - \'prediction_text\': list of possible texts for the answer, as a list of strings depending on a threshold on the confidence probability of each prediction. references: List of question-answers dictionaries with the following key-values: - \'id\': id of the question-answer pair (see above), - \'answers\': a Dict in the CUAD dataset format { \'text\': list of possible texts for the answer, as a list of strings \'answer_start\': list of start positions for the answer, as a list of ints } Note that answer_start values are not taken into account to compute the metric. Returns: \'exact_match\': Exact match (the normalized answer exactly match the gold answer) \'f1\': The F-score of predicted tokens versus the gold answer \'aupr\': Area Under the Precision-Recall curve \'prec_at_80_recall\': Precision at 80% recall \'prec_at_90_recall\': Precision at 90% recall Examples: >>> predictions = [{\'prediction_text\': [\'The seller:\', \'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.\'], \'id\': \'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties\'}] >>> references = [{\'answers\': {\'answer_start\': [143, 49], \'text\': [\'The seller:\', \'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.\']}, \'id\': \'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties\'}] >>> cuad_metric = datasets.load_metric("cuad") >>> results = cuad_metric.compute(predictions=predictions, references=references) >>> print(results) {\'exact_match\': 100.0, \'f1\': 100.0, \'aupr\': 0.0, \'prec_at_80_recall\': 1.0, \'prec_at_90_recall\': 1.0} ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a_ ( datasets.Metric ): """simple docstring""" def __lowerCAmelCase ( self ) ->Tuple: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': { '''id''': datasets.Value('''string''' ), '''prediction_text''': datasets.features.Sequence(datasets.Value('''string''' ) ), }, '''references''': { '''id''': datasets.Value('''string''' ), '''answers''': datasets.features.Sequence( { '''text''': datasets.Value('''string''' ), '''answer_start''': datasets.Value('''int32''' ), } ), }, } ) , codebase_urls=['''https://www.atticusprojectai.org/cuad'''] , reference_urls=['''https://www.atticusprojectai.org/cuad'''] , ) def __lowerCAmelCase ( self , _lowerCamelCase , _lowerCamelCase ) ->Tuple: SCREAMING_SNAKE_CASE : Any = {prediction['''id''']: prediction['''prediction_text'''] for prediction in predictions} SCREAMING_SNAKE_CASE : int = [ { '''paragraphs''': [ { '''qas''': [ { '''answers''': [{'''text''': answer_text} for answer_text in ref['''answers''']['''text''']], '''id''': ref['''id'''], } for ref in references ] } ] } ] SCREAMING_SNAKE_CASE : Dict = evaluate(dataset=_lowerCamelCase , predictions=_lowerCamelCase ) return score
19
1
from __future__ import annotations from collections.abc import Iterable, Iterator from dataclasses import dataclass A : Tuple = (3, 9, -1_1, 0, 7, 5, 1, -1) A : int = (4, 6, 2, 0, 8, 1_0, 3, -2) @dataclass class A : '''simple docstring''' __lowerCamelCase : int __lowerCamelCase : Node | None class A : '''simple docstring''' def __init__( self : Optional[Any] , __lowerCAmelCase : Iterable[int] ) -> None: """simple docstring""" A__ = None for i in sorted(__lowerCAmelCase , reverse=__lowerCAmelCase ): A__ = Node(__lowerCAmelCase , self.head ) def __iter__( self : Dict ) -> Iterator[int]: """simple docstring""" A__ = self.head while node: yield node.data A__ = node.next_node def __len__( self : Union[str, Any] ) -> int: """simple docstring""" return sum(1 for _ in self ) def __str__( self : str ) -> str: """simple docstring""" return " -> ".join([str(__lowerCAmelCase ) for node in self] ) def __lowerCamelCase ( __a :SortedLinkedList , __a :SortedLinkedList ) -> SortedLinkedList: """simple docstring""" return SortedLinkedList(list(__a ) + list(__a ) ) if __name__ == "__main__": import doctest doctest.testmod() A : Optional[int] = SortedLinkedList print(merge_lists(SSL(test_data_odd), SSL(test_data_even)))
274
from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean A : str = 0 A : Any = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] A : Dict = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right A : Union[str, Any] = tuple[int, int] class A : '''simple docstring''' def __init__( self : int , __lowerCAmelCase : int , __lowerCAmelCase : int , __lowerCAmelCase : int , __lowerCAmelCase : int , __lowerCAmelCase : int , __lowerCAmelCase : Node | None , ) -> None: """simple docstring""" A__ = pos_x A__ = pos_y A__ = (pos_y, pos_x) A__ = goal_x A__ = goal_y A__ = g_cost A__ = parent A__ = self.calculate_heuristic() A__ = self.g_cost + self.h_cost def a_ ( self : Dict ) -> float: """simple docstring""" A__ = self.pos_x - self.goal_x A__ = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(__lowerCAmelCase ) + abs(__lowerCAmelCase ) else: return sqrt(dy**2 + dx**2 ) def __lt__( self : int , __lowerCAmelCase : Node ) -> bool: """simple docstring""" return self.f_cost < other.f_cost class A : '''simple docstring''' def __init__( self : Union[str, Any] , __lowerCAmelCase : TPosition , __lowerCAmelCase : TPosition ) -> Tuple: """simple docstring""" A__ = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , __lowerCAmelCase ) A__ = Node(goal[1] , goal[0] , goal[1] , goal[0] , 9_99_99 , __lowerCAmelCase ) A__ = [self.start] A__ = [] A__ = False def a_ ( self : List[str] ) -> list[TPosition]: """simple docstring""" while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() A__ = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: return self.retrace_path(__lowerCAmelCase ) self.closed_nodes.append(__lowerCAmelCase ) A__ = self.get_successors(__lowerCAmelCase ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(__lowerCAmelCase ) else: # retrieve the best current path A__ = self.open_nodes.pop(self.open_nodes.index(__lowerCAmelCase ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(__lowerCAmelCase ) else: self.open_nodes.append(__lowerCAmelCase ) return [self.start.pos] def a_ ( self : Optional[Any] , __lowerCAmelCase : Node ) -> list[Node]: """simple docstring""" A__ = [] for action in delta: A__ = parent.pos_x + action[1] A__ = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(__lowerCAmelCase ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( __lowerCAmelCase , __lowerCAmelCase , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , __lowerCAmelCase , ) ) return successors def a_ ( self : List[Any] , __lowerCAmelCase : Node | None ) -> list[TPosition]: """simple docstring""" A__ = node A__ = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) A__ = current_node.parent path.reverse() return path class A : '''simple docstring''' def __init__( self : Optional[Any] , __lowerCAmelCase : TPosition , __lowerCAmelCase : TPosition ) -> None: """simple docstring""" A__ = AStar(__lowerCAmelCase , __lowerCAmelCase ) A__ = AStar(__lowerCAmelCase , __lowerCAmelCase ) A__ = False def a_ ( self : int ) -> list[TPosition]: """simple docstring""" while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() A__ = self.fwd_astar.open_nodes.pop(0 ) A__ = self.bwd_astar.open_nodes.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( __lowerCAmelCase , __lowerCAmelCase ) self.fwd_astar.closed_nodes.append(__lowerCAmelCase ) self.bwd_astar.closed_nodes.append(__lowerCAmelCase ) A__ = current_bwd_node A__ = current_fwd_node A__ = { self.fwd_astar: self.fwd_astar.get_successors(__lowerCAmelCase ), self.bwd_astar: self.bwd_astar.get_successors(__lowerCAmelCase ), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(__lowerCAmelCase ) else: # retrieve the best current path A__ = astar.open_nodes.pop( astar.open_nodes.index(__lowerCAmelCase ) ) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(__lowerCAmelCase ) else: astar.open_nodes.append(__lowerCAmelCase ) return [self.fwd_astar.start.pos] def a_ ( self : List[str] , __lowerCAmelCase : Node , __lowerCAmelCase : Node ) -> list[TPosition]: """simple docstring""" A__ = self.fwd_astar.retrace_path(__lowerCAmelCase ) A__ = self.bwd_astar.retrace_path(__lowerCAmelCase ) bwd_path.pop() bwd_path.reverse() A__ = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] A : Optional[int] = (0, 0) A : int = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) A : Dict = time.time() A : Optional[Any] = AStar(init, goal) A : Optional[int] = a_star.search() A : Optional[int] = time.time() - start_time print(F'''AStar execution time = {end_time:f} seconds''') A : Dict = time.time() A : Tuple = BidirectionalAStar(init, goal) A : List[Any] = time.time() - bd_start_time print(F'''BidirectionalAStar execution time = {bd_end_time:f} seconds''')
274
1
import bza import gzip import lzma import os import shutil import struct import tarfile import warnings import zipfile from abc import ABC, abstractmethod from pathlib import Path from typing import Dict, List, Optional, Type, Union from .. import config from .filelock import FileLock from .logging import get_logger _UpperCAmelCase : Optional[int] = get_logger(__name__) class lowercase : def __init__( self , snake_case = None ): snake_case_ = ( os.path.join(snake_case , config.EXTRACTED_DATASETS_DIR ) if cache_dir else config.EXTRACTED_DATASETS_PATH ) snake_case_ = Extractor def a ( self , snake_case ): from .file_utils import hash_url_to_filename # Path where we extract compressed archives # We extract in the cache dir, and get the extracted path name by hashing the original path" snake_case_ = os.path.abspath(snake_case ) return os.path.join(self.extract_dir , hash_url_to_filename(snake_case ) ) def a ( self , snake_case , snake_case ): return force_extract or ( not os.path.isfile(snake_case ) and not (os.path.isdir(snake_case ) and os.listdir(snake_case )) ) def a ( self , snake_case , snake_case = False ): snake_case_ = self.extractor.infer_extractor_format(snake_case ) if not extractor_format: return input_path snake_case_ = self._get_output_path(snake_case ) if self._do_extract(snake_case , snake_case ): self.extractor.extract(snake_case , snake_case , snake_case ) return output_path class lowercase ( lowercase_ ): @classmethod @abstractmethod def a ( cls , snake_case , **snake_case ): ... @staticmethod @abstractmethod def a ( snake_case , snake_case ): ... class lowercase ( lowercase_ , lowercase_ ): __SCREAMING_SNAKE_CASE : List[bytes] = [] @staticmethod def a ( snake_case , snake_case ): with open(snake_case , 'rb' ) as f: return f.read(snake_case ) @classmethod def a ( cls , snake_case , snake_case = b"" ): if not magic_number: snake_case_ = max(len(snake_case ) for cls_magic_number in cls.magic_numbers ) try: snake_case_ = cls.read_magic_number(snake_case , snake_case ) except OSError: return False return any(magic_number.startswith(snake_case ) for cls_magic_number in cls.magic_numbers ) class lowercase ( lowercase_ ): @classmethod def a ( cls , snake_case , **snake_case ): return tarfile.is_tarfile(snake_case ) @staticmethod def a ( snake_case , snake_case ): def resolved(snake_case ) -> str: return os.path.realpath(os.path.abspath(snake_case ) ) def badpath(snake_case , snake_case ) -> bool: # joinpath will ignore base if path is absolute return not resolved(os.path.join(snake_case , snake_case ) ).startswith(snake_case ) def badlink(snake_case , snake_case ) -> bool: # Links are interpreted relative to the directory containing the link snake_case_ = resolved(os.path.join(snake_case , os.path.dirname(info.name ) ) ) return badpath(info.linkname , base=snake_case ) snake_case_ = resolved(snake_case ) for finfo in members: if badpath(finfo.name , snake_case ): logger.error(F'''Extraction of {finfo.name} is blocked (illegal path)''' ) elif finfo.issym() and badlink(snake_case , snake_case ): logger.error(F'''Extraction of {finfo.name} is blocked: Symlink to {finfo.linkname}''' ) elif finfo.islnk() and badlink(snake_case , snake_case ): logger.error(F'''Extraction of {finfo.name} is blocked: Hard link to {finfo.linkname}''' ) else: yield finfo @staticmethod def a ( snake_case , snake_case ): os.makedirs(snake_case , exist_ok=snake_case ) snake_case_ = tarfile.open(snake_case ) tar_file.extractall(snake_case , members=TarExtractor.safemembers(snake_case , snake_case ) ) tar_file.close() class lowercase ( lowercase_ ): __SCREAMING_SNAKE_CASE : Optional[int] = [b'''\x1F\x8B'''] @staticmethod def a ( snake_case , snake_case ): with gzip.open(snake_case , 'rb' ) as gzip_file: with open(snake_case , 'wb' ) as extracted_file: shutil.copyfileobj(snake_case , snake_case ) class lowercase ( lowercase_ ): __SCREAMING_SNAKE_CASE : Tuple = [ b'''PK\x03\x04''', b'''PK\x05\x06''', # empty archive b'''PK\x07\x08''', # spanned archive ] @classmethod def a ( cls , snake_case , snake_case = b"" ): if super().is_extractable(snake_case , magic_number=snake_case ): return True try: # Alternative version of zipfile.is_zipfile that has less false positives, but misses executable zip archives. # From: https://github.com/python/cpython/pull/5053 from zipfile import ( _CD_SIGNATURE, _ECD_DISK_NUMBER, _ECD_DISK_START, _ECD_ENTRIES_TOTAL, _ECD_OFFSET, _ECD_SIZE, _EndRecData, sizeCentralDir, stringCentralDir, structCentralDir, ) with open(snake_case , 'rb' ) as fp: snake_case_ = _EndRecData(snake_case ) if endrec: if endrec[_ECD_ENTRIES_TOTAL] == 0 and endrec[_ECD_SIZE] == 0 and endrec[_ECD_OFFSET] == 0: return True # Empty zipfiles are still zipfiles elif endrec[_ECD_DISK_NUMBER] == endrec[_ECD_DISK_START]: fp.seek(endrec[_ECD_OFFSET] ) # Central directory is on the same disk if fp.tell() == endrec[_ECD_OFFSET] and endrec[_ECD_SIZE] >= sizeCentralDir: snake_case_ = fp.read(snake_case ) # CD is where we expect it to be if len(snake_case ) == sizeCentralDir: snake_case_ = struct.unpack(snake_case , snake_case ) # CD is the right size if centdir[_CD_SIGNATURE] == stringCentralDir: return True # First central directory entry has correct magic number return False except Exception: # catch all errors in case future python versions change the zipfile internals return False @staticmethod def a ( snake_case , snake_case ): os.makedirs(snake_case , exist_ok=snake_case ) with zipfile.ZipFile(snake_case , 'r' ) as zip_file: zip_file.extractall(snake_case ) zip_file.close() class lowercase ( lowercase_ ): __SCREAMING_SNAKE_CASE : str = [b'''\xFD\x37\x7A\x58\x5A\x00'''] @staticmethod def a ( snake_case , snake_case ): with lzma.open(snake_case ) as compressed_file: with open(snake_case , 'wb' ) as extracted_file: shutil.copyfileobj(snake_case , snake_case ) class lowercase ( lowercase_ ): __SCREAMING_SNAKE_CASE : List[str] = [b'''Rar!\x1a\x07\x00''', b'''Rar!\x1a\x07\x01\x00'''] # RAR_ID # RAR5_ID @staticmethod def a ( snake_case , snake_case ): if not config.RARFILE_AVAILABLE: raise ImportError('Please pip install rarfile' ) import rarfile os.makedirs(snake_case , exist_ok=snake_case ) snake_case_ = rarfile.RarFile(snake_case ) rf.extractall(snake_case ) rf.close() class lowercase ( lowercase_ ): __SCREAMING_SNAKE_CASE : List[str] = [b'''\x28\xb5\x2F\xFD'''] @staticmethod def a ( snake_case , snake_case ): if not config.ZSTANDARD_AVAILABLE: raise ImportError('Please pip install zstandard' ) import zstandard as zstd snake_case_ = zstd.ZstdDecompressor() with open(snake_case , 'rb' ) as ifh, open(snake_case , 'wb' ) as ofh: dctx.copy_stream(snake_case , snake_case ) class lowercase ( lowercase_ ): __SCREAMING_SNAKE_CASE : Union[str, Any] = [b'''\x42\x5A\x68'''] @staticmethod def a ( snake_case , snake_case ): with bza.open(snake_case , 'rb' ) as compressed_file: with open(snake_case , 'wb' ) as extracted_file: shutil.copyfileobj(snake_case , snake_case ) class lowercase ( lowercase_ ): __SCREAMING_SNAKE_CASE : Union[str, Any] = [b'''\x37\x7A\xBC\xAF\x27\x1C'''] @staticmethod def a ( snake_case , snake_case ): if not config.PY7ZR_AVAILABLE: raise ImportError('Please pip install py7zr' ) import pyazr os.makedirs(snake_case , exist_ok=snake_case ) with pyazr.SevenZipFile(snake_case , 'r' ) as archive: archive.extractall(snake_case ) class lowercase ( lowercase_ ): __SCREAMING_SNAKE_CASE : Any = [b'''\x04\x22\x4D\x18'''] @staticmethod def a ( snake_case , snake_case ): if not config.LZ4_AVAILABLE: raise ImportError('Please pip install lz4' ) import lza.frame with lza.frame.open(snake_case , 'rb' ) as compressed_file: with open(snake_case , 'wb' ) as extracted_file: shutil.copyfileobj(snake_case , snake_case ) class lowercase : # Put zip file to the last, b/c it is possible wrongly detected as zip (I guess it means: as tar or gzip) __SCREAMING_SNAKE_CASE : Dict[str, Type[BaseExtractor]] = { "tar": TarExtractor, "gzip": GzipExtractor, "zip": ZipExtractor, "xz": XzExtractor, "rar": RarExtractor, "zstd": ZstdExtractor, "bz2": BzipaExtractor, "7z": SevenZipExtractor, # <Added version="2.4.0"/> "lz4": LzaExtractor, # <Added version="2.4.0"/> } @classmethod def a ( cls ): return max( len(snake_case ) for extractor in cls.extractors.values() if issubclass(snake_case , snake_case ) for extractor_magic_number in extractor.magic_numbers ) @staticmethod def a ( snake_case , snake_case ): try: return MagicNumberBaseExtractor.read_magic_number(snake_case , magic_number_length=snake_case ) except OSError: return b"" @classmethod def a ( cls , snake_case , snake_case = False ): warnings.warn( 'Method \'is_extractable\' was deprecated in version 2.4.0 and will be removed in 3.0.0. ' 'Use \'infer_extractor_format\' instead.' , category=snake_case , ) snake_case_ = cls.infer_extractor_format(snake_case ) if extractor_format: return True if not return_extractor else (True, cls.extractors[extractor_format]) return False if not return_extractor else (False, None) @classmethod def a ( cls , snake_case ): # <Added version="2.4.0"/> snake_case_ = cls._get_magic_number_max_length() snake_case_ = cls._read_magic_number(snake_case , snake_case ) for extractor_format, extractor in cls.extractors.items(): if extractor.is_extractable(snake_case , magic_number=snake_case ): return extractor_format @classmethod def a ( cls , snake_case , snake_case , snake_case = None , snake_case = "deprecated" , ): os.makedirs(os.path.dirname(snake_case ) , exist_ok=snake_case ) # Prevent parallel extractions snake_case_ = str(Path(snake_case ).with_suffix('.lock' ) ) with FileLock(snake_case ): shutil.rmtree(snake_case , ignore_errors=snake_case ) if extractor_format or extractor != "deprecated": if extractor != "deprecated" or not isinstance(snake_case , snake_case ): # passed as positional arg warnings.warn( 'Parameter \'extractor\' was deprecated in version 2.4.0 and will be removed in 3.0.0. ' 'Use \'extractor_format\' instead.' , category=snake_case , ) snake_case_ = extractor if extractor != 'deprecated' else extractor_format else: snake_case_ = cls.extractors[extractor_format] return extractor.extract(snake_case , snake_case ) else: warnings.warn( 'Parameter \'extractor_format\' was made required in version 2.4.0 and not passing it will raise an ' 'exception in 3.0.0.' , category=snake_case , ) for extractor in cls.extractors.values(): if extractor.is_extractable(snake_case ): return extractor.extract(snake_case , snake_case )
200
def __lowerCamelCase ( UpperCamelCase__ ): '''simple docstring''' snake_case_ = 1 snake_case_ = 2 while i * i <= n: snake_case_ = 0 while n % i == 0: n //= i multiplicity += 1 n_divisors *= multiplicity + 1 i += 1 if n > 1: n_divisors *= 2 return n_divisors def __lowerCamelCase ( ): '''simple docstring''' snake_case_ = 1 snake_case_ = 1 while True: i += 1 t_num += i if count_divisors(UpperCamelCase__ ) > 500: break return t_num if __name__ == "__main__": print(solution())
200
1
from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class UpperCAmelCase_ ( _UpperCamelCase ): '''simple docstring''' def __init__( self , __A , __A = None , __A = None , __A = False , __A = False , __A = None , __A = None , **__A , ): """simple docstring""" super().__init__( features=__lowerCamelCase , cache_dir=__lowerCamelCase , keep_in_memory=__lowerCamelCase , streaming=__lowerCamelCase , num_proc=__lowerCamelCase , **__lowerCamelCase , ) lowerCamelCase : str = Generator( cache_dir=__lowerCamelCase , features=__lowerCamelCase , generator=__lowerCamelCase , gen_kwargs=__lowerCamelCase , **__lowerCamelCase , ) def _snake_case ( self ): """simple docstring""" if self.streaming: lowerCamelCase : Tuple = self.builder.as_streaming_dataset(split="train" ) # Build regular (map-style) dataset else: lowerCamelCase : int = None lowerCamelCase : Optional[int] = None lowerCamelCase : Tuple = None lowerCamelCase : str = None self.builder.download_and_prepare( download_config=__lowerCamelCase , download_mode=__lowerCamelCase , verification_mode=__lowerCamelCase , base_path=__lowerCamelCase , num_proc=self.num_proc , ) lowerCamelCase : Any = self.builder.as_dataset( split="train" , verification_mode=__lowerCamelCase , in_memory=self.keep_in_memory ) return dataset
283
import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __lowerCAmelCase : str = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class snake_case__ (_UpperCamelCase , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = XLMRobertaTokenizer SCREAMING_SNAKE_CASE_ : int = XLMRobertaTokenizerFast SCREAMING_SNAKE_CASE_ : str = True SCREAMING_SNAKE_CASE_ : Optional[Any] = True def __UpperCAmelCase ( self : int ) -> List[str]: super().setUp() # We have a SentencePiece fixture for testing a = XLMRobertaTokenizer(__lowerCamelCase , keep_accents=__lowerCamelCase ) tokenizer.save_pretrained(self.tmpdirname ) def __UpperCAmelCase ( self : List[str] ) -> Any: a = "<pad>" a = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__lowerCamelCase ) , __lowerCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__lowerCamelCase ) , __lowerCamelCase ) def __UpperCAmelCase ( self : Dict ) -> Optional[Any]: a = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<s>" ) self.assertEqual(vocab_keys[1] , "<pad>" ) self.assertEqual(vocab_keys[-1] , "<mask>" ) self.assertEqual(len(__lowerCamelCase ) , 10_02 ) def __UpperCAmelCase ( self : List[Any] ) -> Any: self.assertEqual(self.get_tokenizer().vocab_size , 10_02 ) def __UpperCAmelCase ( self : Dict ) -> List[str]: a = XLMRobertaTokenizer(__lowerCamelCase , keep_accents=__lowerCamelCase ) a = tokenizer.tokenize("This is a test" ) self.assertListEqual(__lowerCamelCase , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , [value + tokenizer.fairseq_offset for value in [2_85, 46, 10, 1_70, 3_82]] , ) a = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) a = tokenizer.convert_tokens_to_ids(__lowerCamelCase ) self.assertListEqual( __lowerCamelCase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) a = tokenizer.convert_ids_to_tokens(__lowerCamelCase ) self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) def __UpperCAmelCase ( self : Optional[Any] ) -> List[str]: if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return a = (self.rust_tokenizer_class, "hf-internal-testing/tiny-xlm-roberta", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): a = self.rust_tokenizer_class.from_pretrained(__lowerCamelCase , **__lowerCamelCase ) a = self.tokenizer_class.from_pretrained(__lowerCamelCase , **__lowerCamelCase ) a = tempfile.mkdtemp() a = tokenizer_r.save_pretrained(__lowerCamelCase ) a = tokenizer_p.save_pretrained(__lowerCamelCase ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files ) ) a = tuple(f for f in tokenizer_r_files if "tokenizer.json" not in f ) self.assertSequenceEqual(__lowerCamelCase , __lowerCamelCase ) # Checks everything loads correctly in the same way a = tokenizer_r.from_pretrained(__lowerCamelCase ) a = tokenizer_p.from_pretrained(__lowerCamelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__lowerCamelCase , __lowerCamelCase ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(__lowerCamelCase ) # Save tokenizer rust, legacy_format=True a = tempfile.mkdtemp() a = tokenizer_r.save_pretrained(__lowerCamelCase , legacy_format=__lowerCamelCase ) a = tokenizer_p.save_pretrained(__lowerCamelCase ) # Checks it save with the same files self.assertSequenceEqual(__lowerCamelCase , __lowerCamelCase ) # Checks everything loads correctly in the same way a = tokenizer_r.from_pretrained(__lowerCamelCase ) a = tokenizer_p.from_pretrained(__lowerCamelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__lowerCamelCase , __lowerCamelCase ) ) shutil.rmtree(__lowerCamelCase ) # Save tokenizer rust, legacy_format=False a = tempfile.mkdtemp() a = tokenizer_r.save_pretrained(__lowerCamelCase , legacy_format=__lowerCamelCase ) a = tokenizer_p.save_pretrained(__lowerCamelCase ) # Checks it saved the tokenizer.json file self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way a = tokenizer_r.from_pretrained(__lowerCamelCase ) a = tokenizer_p.from_pretrained(__lowerCamelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__lowerCamelCase , __lowerCamelCase ) ) shutil.rmtree(__lowerCamelCase ) @cached_property def __UpperCAmelCase ( self : List[str] ) -> Union[str, Any]: return XLMRobertaTokenizer.from_pretrained("xlm-roberta-base" ) def __UpperCAmelCase ( self : List[Any] ) -> List[Any]: with tempfile.NamedTemporaryFile() as f: shutil.copyfile(__lowerCamelCase , f.name ) a = XLMRobertaTokenizer(f.name , keep_accents=__lowerCamelCase ) a = pickle.dumps(__lowerCamelCase ) pickle.loads(__lowerCamelCase ) def __UpperCAmelCase ( self : int ) -> str: if not self.test_rust_tokenizer: return a = self.get_tokenizer() a = self.get_rust_tokenizer() a = "I was born in 92000, and this is falsé." a = tokenizer.tokenize(__lowerCamelCase ) a = rust_tokenizer.tokenize(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) a = tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) a = rust_tokenizer.encode(__lowerCamelCase , add_special_tokens=__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) a = self.get_rust_tokenizer() a = tokenizer.encode(__lowerCamelCase ) a = rust_tokenizer.encode(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) @slow def __UpperCAmelCase ( self : Dict ) -> Any: a = "Hello World!" a = [0, 3_53_78, 66_61, 38, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(__lowerCamelCase , self.big_tokenizer.encode(__lowerCamelCase ) ) @slow def __UpperCAmelCase ( self : Tuple ) -> int: a = ( "This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will" " add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth" ) a = [ 0, 32_93, 83, 10, 45_52, 49_89, 79_86, 6_78, 10, 59_15, 1_11, 17_94_59, 12_48_50, 4, 60_44, 2_37, 12, 6, 5, 6, 4, 67_80, 7_05, 15, 13_88, 44, 3_78, 1_01_14, 7_11, 1_52, 20, 6, 5, 2_23_76, 6_42, 12_21, 1_51_90, 3_41_53, 4_50, 56_08, 9_59, 11_19, 5_77_02, 1_36, 1_86, 47, 10_98, 2_93_67, 47, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 60_44, 2_37, 62_84, 5_09_01, 5_28, 31, 90, 34, 9_27, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(__lowerCamelCase , self.big_tokenizer.encode(__lowerCamelCase ) ) @slow def __UpperCAmelCase ( self : Optional[Any] ) -> List[Any]: # fmt: off a = {"input_ids": [[0, 1_10_62, 8_27_72, 7, 15, 8_27_72, 5_38, 5_15_29, 2_37, 1_71_98, 12_90, 2_06, 9, 21_51_75, 13_14, 1_36, 1_71_98, 12_90, 2_06, 9, 5_63_59, 42, 12_20_09, 9, 1_64_66, 16, 8_73_44, 45_37, 9, 47_17, 7_83_81, 6, 15_99_58, 7, 15, 2_44_80, 6_18, 4, 5_27, 2_26_93, 54_28, 4, 27_77, 2_44_80, 98_74, 4, 4_35_23, 5_94, 4, 8_03, 1_83_92, 3_31_89, 18, 4, 4_35_23, 2_44_47, 1_23_99, 1_00, 2_49_55, 8_36_58, 96_26, 14_40_57, 15, 8_39, 2_23_35, 16, 1_36, 2_49_55, 8_36_58, 8_34_79, 15, 3_91_02, 7_24, 16, 6_78, 6_45, 27_89, 13_28, 45_89, 42, 12_20_09, 11_57_74, 23, 8_05, 13_28, 4_68_76, 7, 1_36, 5_38_94, 19_40, 4_22_27, 4_11_59, 1_77_21, 8_23, 4_25, 4, 2_75_12, 9_87_22, 2_06, 1_36, 55_31, 49_70, 9_19, 1_73_36, 5, 2], [0, 2_00_80, 6_18, 83, 8_27_75, 47, 4_79, 9, 15_17, 73, 5_38_94, 3_33, 8_05_81, 11_01_17, 1_88_11, 52_56, 12_95, 51, 15_25_26, 2_97, 79_86, 3_90, 12_44_16, 5_38, 3_54_31, 2_14, 98, 1_50_44, 2_57_37, 1_36, 71_08, 4_37_01, 23, 7_56, 13_53_55, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 5_81, 6_37_73, 11_94_55, 6, 14_77_97, 8_82_03, 7, 6_45, 70, 21, 32_85, 1_02_69, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__lowerCamelCase , model_name="xlm-roberta-base" , revision="d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3" , )
107
0
'''simple docstring''' import dataclasses import json import warnings from dataclasses import dataclass, field from time import time from typing import List from ..utils import logging _lowercase : Tuple = logging.get_logger(__name__) def lowerCamelCase__ ( A : Optional[Any]=None , A : Optional[Any]=None ): '''simple docstring''' return field(default_factory=lambda: default , metadata=lowercase__ ) @dataclass class UpperCamelCase__: __magic_name__ : List[str] = list_field( default=[] , metadata={ "help": ( "Model checkpoints to be provided to the AutoModel classes. Leave blank to benchmark the base version" " of all available models" ) } , ) __magic_name__ : List[int] = list_field( default=[8] , metadata={"help": "List of batch sizes for which memory and time performance will be evaluated"} ) __magic_name__ : List[int] = list_field( default=[8, 32, 128, 512] , metadata={"help": "List of sequence lengths for which memory and time performance will be evaluated"} , ) __magic_name__ : bool = field( default=__UpperCamelCase , metadata={"help": "Whether to benchmark inference of model. Inference can be disabled via --no-inference."} , ) __magic_name__ : bool = field( default=__UpperCamelCase , metadata={"help": "Whether to run on available cuda devices. Cuda can be disabled via --no-cuda."} , ) __magic_name__ : bool = field( default=__UpperCamelCase , metadata={"help": "Whether to run on available tpu devices. TPU can be disabled via --no-tpu."} ) __magic_name__ : bool = field(default=__UpperCamelCase , metadata={"help": "Use FP16 to accelerate inference."} ) __magic_name__ : bool = field(default=__UpperCamelCase , metadata={"help": "Benchmark training of model"} ) __magic_name__ : bool = field(default=__UpperCamelCase , metadata={"help": "Verbose memory tracing"} ) __magic_name__ : bool = field( default=__UpperCamelCase , metadata={"help": "Whether to perform speed measurements. Speed measurements can be disabled via --no-speed."} , ) __magic_name__ : bool = field( default=__UpperCamelCase , metadata={ "help": "Whether to perform memory measurements. Memory measurements can be disabled via --no-memory" } , ) __magic_name__ : bool = field(default=__UpperCamelCase , metadata={"help": "Trace memory line by line"} ) __magic_name__ : bool = field(default=__UpperCamelCase , metadata={"help": "Save result to a CSV file"} ) __magic_name__ : bool = field(default=__UpperCamelCase , metadata={"help": "Save all print statements in a log file"} ) __magic_name__ : bool = field(default=__UpperCamelCase , metadata={"help": "Whether to print environment information"} ) __magic_name__ : bool = field( default=__UpperCamelCase , metadata={ "help": ( "Whether to use multiprocessing for memory and speed measurement. It is highly recommended to use" " multiprocessing for accurate CPU and GPU memory measurements. This option should only be disabled" " for debugging / testing and on TPU." ) } , ) __magic_name__ : str = field( default=f'inference_time_{round(time() )}.csv' , metadata={"help": "CSV filename used if saving time results to csv."} , ) __magic_name__ : str = field( default=f'inference_memory_{round(time() )}.csv' , metadata={"help": "CSV filename used if saving memory results to csv."} , ) __magic_name__ : str = field( default=f'train_time_{round(time() )}.csv' , metadata={"help": "CSV filename used if saving time results to csv for training."} , ) __magic_name__ : str = field( default=f'train_memory_{round(time() )}.csv' , metadata={"help": "CSV filename used if saving memory results to csv for training."} , ) __magic_name__ : str = field( default=f'env_info_{round(time() )}.csv' , metadata={"help": "CSV filename used if saving environment information."} , ) __magic_name__ : str = field( default=f'log_{round(time() )}.csv' , metadata={"help": "Log filename used if print statements are saved in log."} , ) __magic_name__ : int = field(default=3 , metadata={"help": "Times an experiment will be run."} ) __magic_name__ : bool = field( default=__UpperCamelCase , metadata={ "help": ( "Instead of loading the model as defined in `config.architectures` if exists, just load the pretrain" " model weights." ) } , ) def a__( self : Any )-> List[str]: """simple docstring""" warnings.warn( F"""The class {self.__class__} is deprecated. Hugging Face Benchmarking utils""" ''' are deprecated in general and it is advised to use external Benchmarking libraries ''' ''' to benchmark Transformer models.''' , lowerCAmelCase , ) def a__( self : Dict )-> Optional[int]: """simple docstring""" return json.dumps(dataclasses.asdict(self ) , indent=2 ) @property def a__( self : Union[str, Any] )-> List[Any]: """simple docstring""" if len(self.models ) <= 0: raise ValueError( '''Please make sure you provide at least one model name / model identifier, *e.g.* `--models''' ''' bert-base-cased` or `args.models = [\'bert-base-cased\'].''' ) return self.models @property def a__( self : Union[str, Any] )-> Dict: """simple docstring""" if not self.multi_process: return False elif self.is_tpu: logger.info('''Multiprocessing is currently not possible on TPU.''' ) return False else: return True
356
'''simple docstring''' import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class UpperCamelCase__( lowerCAmelCase ): __magic_name__ : List[Any] = ["image_processor", "tokenizer"] __magic_name__ : Tuple = "ViTImageProcessor" __magic_name__ : int = ("CLIPTokenizer", "CLIPTokenizerFast") def __init__( self : List[str] , lowerCAmelCase : Tuple=None , lowerCAmelCase : List[str]=None , **lowerCAmelCase : Optional[int] )-> Tuple: """simple docstring""" UpperCAmelCase = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , lowerCAmelCase , ) UpperCAmelCase = kwargs.pop('''feature_extractor''' ) UpperCAmelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(lowerCAmelCase , lowerCAmelCase ) def __call__( self : int , lowerCAmelCase : Dict=None , lowerCAmelCase : Optional[int]=None , lowerCAmelCase : Union[str, Any]=None , lowerCAmelCase : int=None , **lowerCAmelCase : Tuple )-> Optional[int]: """simple docstring""" if text is None and visual_prompt is None and images is None: raise ValueError('''You have to specify either text, visual prompt or images.''' ) if text is not None and visual_prompt is not None: raise ValueError('''You have to specify exactly one type of prompt. Either text or visual prompt.''' ) if text is not None: UpperCAmelCase = self.tokenizer(lowerCAmelCase , return_tensors=lowerCAmelCase , **lowerCAmelCase ) if visual_prompt is not None: UpperCAmelCase = self.image_processor(lowerCAmelCase , return_tensors=lowerCAmelCase , **lowerCAmelCase ) if images is not None: UpperCAmelCase = self.image_processor(lowerCAmelCase , return_tensors=lowerCAmelCase , **lowerCAmelCase ) if visual_prompt is not None and images is not None: UpperCAmelCase = { '''pixel_values''': image_features.pixel_values, '''conditional_pixel_values''': prompt_features.pixel_values, } return encoding elif text is not None and images is not None: UpperCAmelCase = image_features.pixel_values return encoding elif text is not None: return encoding elif visual_prompt is not None: UpperCAmelCase = { '''conditional_pixel_values''': prompt_features.pixel_values, } return encoding else: return BatchEncoding(data=dict(**lowerCAmelCase ) , tensor_type=lowerCAmelCase ) def a__( self : Optional[int] , *lowerCAmelCase : List[str] , **lowerCAmelCase : Dict )-> Tuple: """simple docstring""" return self.tokenizer.batch_decode(*lowerCAmelCase , **lowerCAmelCase ) def a__( self : List[Any] , *lowerCAmelCase : str , **lowerCAmelCase : List[Any] )-> Optional[Any]: """simple docstring""" return self.tokenizer.decode(*lowerCAmelCase , **lowerCAmelCase ) @property def a__( self : Any )-> Optional[int]: """simple docstring""" warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , lowerCAmelCase , ) return self.image_processor_class @property def a__( self : str )-> List[Any]: """simple docstring""" warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , lowerCAmelCase , ) return self.image_processor
91
0
'''simple docstring''' import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("""TEST_SAGEMAKER""" , """False""" ) ) is not True , reason="""Skipping test because should only be run when releasing minor transformers version""" , ) @pytest.mark.usefixtures("""sm_env""" ) @parameterized_class( [ { """framework""": """pytorch""", """script""": """run_glue.py""", """model_name_or_path""": """distilbert-base-cased""", """instance_type""": """ml.g4dn.xlarge""", """results""": {"""train_runtime""": 6_50, """eval_accuracy""": 0.6, """eval_loss""": 0.9}, }, { """framework""": """tensorflow""", """script""": """run_tf.py""", """model_name_or_path""": """distilbert-base-cased""", """instance_type""": """ml.g4dn.xlarge""", """results""": {"""train_runtime""": 6_00, """eval_accuracy""": 0.3, """eval_loss""": 0.9}, }, ] ) class lowercase__ ( unittest.TestCase ): def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' if self.framework == "pytorch": subprocess.run( F'cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py'.split() ,encoding='utf-8' ,check=lowerCamelCase__ ,) assert hasattr(self ,'env' ) def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : Union[str, Any]=1 ): '''simple docstring''' # creates estimator return HuggingFace( entry_point=self.script ,source_dir=self.env.test_path ,role=self.env.role ,image_uri=self.env.image_uri ,base_job_name=F'{self.env.base_job_name}-single' ,instance_count=lowerCamelCase__ ,instance_type=self.instance_type ,debugger_hook_config=lowerCamelCase__ ,hyperparameters={**self.env.hyperparameters, 'model_name_or_path': self.model_name_or_path} ,metric_definitions=self.env.metric_definitions ,py_version='py36' ,) def UpperCamelCase_ ( self : int ,lowerCamelCase__ : Union[str, Any] ): '''simple docstring''' TrainingJobAnalytics(lowerCamelCase__ ).export_csv(F'{self.env.test_path}/{job_name}_metrics.csv' ) def UpperCamelCase_ ( self : int ): '''simple docstring''' # create estimator _UpperCamelCase : Optional[int] = self.create_estimator() # run training estimator.fit() # result dataframe _UpperCamelCase : List[Any] = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis _UpperCamelCase : str = list(result_metrics_df[result_metrics_df.metric_name == 'eval_accuracy']['value'] ) _UpperCamelCase : Optional[int] = list(result_metrics_df[result_metrics_df.metric_name == 'eval_loss']['value'] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping _UpperCamelCase : List[Any] = ( Session().describe_training_job(estimator.latest_training_job.name ).get('TrainingTimeInSeconds' ,999999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results['eval_accuracy'] for t in eval_accuracy ) assert all(t <= self.results['eval_loss'] for t in eval_loss ) # dump tests result into json file to share in PR with open(F'{estimator.latest_training_job.name}.json' ,'w' ) as outfile: json.dump({'train_time': train_runtime, 'eval_accuracy': eval_accuracy, 'eval_loss': eval_loss} ,lowerCamelCase__ )
83
'''simple docstring''' import argparse import os import re _UpperCAmelCase : Tuple = """src/transformers""" # Pattern that looks at the indentation in a line. _UpperCAmelCase : Any = re.compile(r"""^(\s*)\S""") # Pattern that matches `"key":" and puts `key` in group 0. _UpperCAmelCase : List[Any] = re.compile(r"""^\s*\"([^\"]+)\":""") # Pattern that matches `_import_structure["key"]` and puts `key` in group 0. _UpperCAmelCase : Optional[int] = re.compile(r"""^\s*_import_structure\[\"([^\"]+)\"\]""") # Pattern that matches `"key",` and puts `key` in group 0. _UpperCAmelCase : Tuple = re.compile(r"""^\s*\"([^\"]+)\",\s*$""") # Pattern that matches any `[stuff]` and puts `stuff` in group 0. _UpperCAmelCase : Optional[int] = re.compile(r"""\[([^\]]+)\]""") def __magic_name__( lowerCamelCase): __lowerCAmelCase = _re_indent.search(lowerCamelCase) return "" if search is None else search.groups()[0] def __magic_name__( lowerCamelCase, lowerCamelCase="", lowerCamelCase=None, lowerCamelCase=None): __lowerCAmelCase = 0 __lowerCAmelCase = code.split('''\n''') if start_prompt is not None: while not lines[index].startswith(lowerCamelCase): index += 1 __lowerCAmelCase = ['''\n'''.join(lines[:index])] else: __lowerCAmelCase = [] # We split into blocks until we get to the `end_prompt` (or the end of the block). __lowerCAmelCase = [lines[index]] index += 1 while index < len(lowerCamelCase) and (end_prompt is None or not lines[index].startswith(lowerCamelCase)): if len(lines[index]) > 0 and get_indent(lines[index]) == indent_level: if len(lowerCamelCase) > 0 and get_indent(current_block[-1]).startswith(indent_level + ''' '''): current_block.append(lines[index]) blocks.append('''\n'''.join(lowerCamelCase)) if index < len(lowerCamelCase) - 1: __lowerCAmelCase = [lines[index + 1]] index += 1 else: __lowerCAmelCase = [] else: blocks.append('''\n'''.join(lowerCamelCase)) __lowerCAmelCase = [lines[index]] else: current_block.append(lines[index]) index += 1 # Adds current block if it's nonempty. if len(lowerCamelCase) > 0: blocks.append('''\n'''.join(lowerCamelCase)) # Add final block after end_prompt if provided. if end_prompt is not None and index < len(lowerCamelCase): blocks.append('''\n'''.join(lines[index:])) return blocks def __magic_name__( lowerCamelCase): def _inner(lowerCamelCase): return key(lowerCamelCase).lower().replace('''_''', '''''') return _inner def __magic_name__( lowerCamelCase, lowerCamelCase=None): # If no key is provided, we use a noop. def noop(lowerCamelCase): return x if key is None: __lowerCAmelCase = noop # Constants are all uppercase, they go first. __lowerCAmelCase = [obj for obj in objects if key(lowerCamelCase).isupper()] # Classes are not all uppercase but start with a capital, they go second. __lowerCAmelCase = [obj for obj in objects if key(lowerCamelCase)[0].isupper() and not key(lowerCamelCase).isupper()] # Functions begin with a lowercase, they go last. __lowerCAmelCase = [obj for obj in objects if not key(lowerCamelCase)[0].isupper()] __lowerCAmelCase = ignore_underscore(lowerCamelCase) return sorted(lowerCamelCase, key=lowerCamelCase) + sorted(lowerCamelCase, key=lowerCamelCase) + sorted(lowerCamelCase, key=lowerCamelCase) def __magic_name__( lowerCamelCase): # This inner function sort imports between [ ]. def _replace(lowerCamelCase): __lowerCAmelCase = match.groups()[0] if "," not in imports: return F"""[{imports}]""" __lowerCAmelCase = [part.strip().replace('''"''', '''''') for part in imports.split(''',''')] # We will have a final empty element if the line finished with a comma. if len(keys[-1]) == 0: __lowerCAmelCase = keys[:-1] return "[" + ", ".join([F"""\"{k}\"""" for k in sort_objects(lowerCamelCase)]) + "]" __lowerCAmelCase = import_statement.split('''\n''') if len(lowerCamelCase) > 3: # Here we have to sort internal imports that are on several lines (one per name): # key: [ # "object1", # "object2", # ... # ] # We may have to ignore one or two lines on each side. __lowerCAmelCase = 2 if lines[1].strip() == '''[''' else 1 __lowerCAmelCase = [(i, _re_strip_line.search(lowerCamelCase).groups()[0]) for i, line in enumerate(lines[idx:-idx])] __lowerCAmelCase = sort_objects(lowerCamelCase, key=lambda lowerCamelCase: x[1]) __lowerCAmelCase = [lines[x[0] + idx] for x in sorted_indices] return "\n".join(lines[:idx] + sorted_lines + lines[-idx:]) elif len(lowerCamelCase) == 3: # Here we have to sort internal imports that are on one separate line: # key: [ # "object1", "object2", ... # ] if _re_bracket_content.search(lines[1]) is not None: __lowerCAmelCase = _re_bracket_content.sub(_replace, lines[1]) else: __lowerCAmelCase = [part.strip().replace('''"''', '''''') for part in lines[1].split(''',''')] # We will have a final empty element if the line finished with a comma. if len(keys[-1]) == 0: __lowerCAmelCase = keys[:-1] __lowerCAmelCase = get_indent(lines[1]) + ''', '''.join([F"""\"{k}\"""" for k in sort_objects(lowerCamelCase)]) return "\n".join(lowerCamelCase) else: # Finally we have to deal with imports fitting on one line __lowerCAmelCase = _re_bracket_content.sub(_replace, lowerCamelCase) return import_statement def __magic_name__( lowerCamelCase, lowerCamelCase=True): with open(lowerCamelCase, encoding='''utf-8''') as f: __lowerCAmelCase = f.read() if "_import_structure" not in code: return # Blocks of indent level 0 __lowerCAmelCase = split_code_in_indented_blocks( lowerCamelCase, start_prompt='''_import_structure = {''', end_prompt='''if TYPE_CHECKING:''') # We ignore block 0 (everything untils start_prompt) and the last block (everything after end_prompt). for block_idx in range(1, len(lowerCamelCase) - 1): # Check if the block contains some `_import_structure`s thingy to sort. __lowerCAmelCase = main_blocks[block_idx] __lowerCAmelCase = block.split('''\n''') # Get to the start of the imports. __lowerCAmelCase = 0 while line_idx < len(lowerCamelCase) and "_import_structure" not in block_lines[line_idx]: # Skip dummy import blocks if "import dummy" in block_lines[line_idx]: __lowerCAmelCase = len(lowerCamelCase) else: line_idx += 1 if line_idx >= len(lowerCamelCase): continue # Ignore beginning and last line: they don't contain anything. __lowerCAmelCase = '''\n'''.join(block_lines[line_idx:-1]) __lowerCAmelCase = get_indent(block_lines[1]) # Slit the internal block into blocks of indent level 1. __lowerCAmelCase = split_code_in_indented_blocks(lowerCamelCase, indent_level=lowerCamelCase) # We have two categories of import key: list or _import_structure[key].append/extend __lowerCAmelCase = _re_direct_key if '''_import_structure = {''' in block_lines[0] else _re_indirect_key # Grab the keys, but there is a trap: some lines are empty or just comments. __lowerCAmelCase = [(pattern.search(lowerCamelCase).groups()[0] if pattern.search(lowerCamelCase) is not None else None) for b in internal_blocks] # We only sort the lines with a key. __lowerCAmelCase = [(i, key) for i, key in enumerate(lowerCamelCase) if key is not None] __lowerCAmelCase = [x[0] for x in sorted(lowerCamelCase, key=lambda lowerCamelCase: x[1])] # We reorder the blocks by leaving empty lines/comments as they were and reorder the rest. __lowerCAmelCase = 0 __lowerCAmelCase = [] for i in range(len(lowerCamelCase)): if keys[i] is None: reorderded_blocks.append(internal_blocks[i]) else: __lowerCAmelCase = sort_objects_in_import(internal_blocks[sorted_indices[count]]) reorderded_blocks.append(lowerCamelCase) count += 1 # And we put our main block back together with its first and last line. __lowerCAmelCase = '''\n'''.join(block_lines[:line_idx] + reorderded_blocks + [block_lines[-1]]) if code != "\n".join(lowerCamelCase): if check_only: return True else: print(F"""Overwriting {file}.""") with open(lowerCamelCase, '''w''', encoding='''utf-8''') as f: f.write('''\n'''.join(lowerCamelCase)) def __magic_name__( lowerCamelCase=True): __lowerCAmelCase = [] for root, _, files in os.walk(lowerCamelCase): if "__init__.py" in files: __lowerCAmelCase = sort_imports(os.path.join(lowerCamelCase, '''__init__.py'''), check_only=lowerCamelCase) if result: __lowerCAmelCase = [os.path.join(lowerCamelCase, '''__init__.py''')] if len(lowerCamelCase) > 0: raise ValueError(F"""Would overwrite {len(lowerCamelCase)} files, run `make style`.""") if __name__ == "__main__": _UpperCAmelCase : str = argparse.ArgumentParser() parser.add_argument("""--check_only""", action="""store_true""", help="""Whether to only check or fix style.""") _UpperCAmelCase : Optional[int] = parser.parse_args() sort_imports_in_all_inits(check_only=args.check_only)
174
0
"""simple docstring""" import argparse import json import os import torch from torch import nn from transformers import NllbMoeConfig, NllbMoeModel from transformers.modeling_utils import dtype_byte_size from transformers.utils import WEIGHTS_INDEX_NAME, WEIGHTS_NAME def _lowerCAmelCase ( UpperCAmelCase : List[str] ): '''simple docstring''' UpperCamelCase__ : Union[str, Any] =[ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''decoder.output_projection.weight''', '''_float_tensor''', '''encoder.embed_positions._float_tensor''', '''decoder.embed_positions._float_tensor''', ] for k in ignore_keys: state_dict.pop(__a , __a ) def _lowerCAmelCase ( UpperCAmelCase : str ): '''simple docstring''' UpperCamelCase__ : List[str] =emb.weight.shape UpperCamelCase__ : Optional[Any] =nn.Linear(__a , __a , bias=__a ) UpperCamelCase__ : Optional[Any] =emb.weight.data return lin_layer def _lowerCAmelCase ( UpperCAmelCase : Any , UpperCAmelCase : str=None ): '''simple docstring''' UpperCamelCase__ : List[Any] ={} for old_key in state_dict.keys(): UpperCamelCase__ : Dict =old_key if "moe_layer.experts." in key: if expert_idx is not None: UpperCamelCase__ : Optional[Any] =key.replace('''moe_layer.experts.0''' , F'''ffn.experts.expert_{expert_idx}''' ) else: UpperCamelCase__ : Tuple =key.replace('''moe_layer.experts.''' , '''ffn.experts.expert_''' ) if "gate" in key: UpperCamelCase__ : Tuple =key.replace('''.moe_layer.gate.wg''' , '''.ffn.router.classifier''' ) if "fc2" and "experts" not in key: UpperCamelCase__ : Optional[Any] =key.replace('''.fc2.''' , '''.ffn.fc2.''' ) if "fc1" and "experts" not in key: UpperCamelCase__ : str =key.replace('''.fc1.''' , '''.ffn.fc1.''' ) if ".encoder_attn." in key: UpperCamelCase__ : str =key.replace('''.encoder_attn.''' , '''.cross_attention.''' ) if "encoder_attn_layer_norm" in key: UpperCamelCase__ : Dict =key.replace('''encoder_attn_layer_norm''' , '''cross_attention_layer_norm''' ) if "final_layer_norm" in key: UpperCamelCase__ : int =key.replace('''final_layer_norm''' , '''ff_layer_norm''' ) UpperCamelCase__ : Any =state_dict[old_key] return new_dict def _lowerCAmelCase ( UpperCAmelCase : Dict , UpperCAmelCase : Optional[Any] , UpperCAmelCase : str , UpperCAmelCase : Dict , UpperCAmelCase : str = WEIGHTS_NAME ): '''simple docstring''' UpperCamelCase__ : List[Any] =[] UpperCamelCase__ : str =0 os.makedirs(__a , exist_ok=__a ) for expert in range(__a ): UpperCamelCase__ : Dict =switch_checkpoint_path + F'''-rank-{expert}.pt''' if os.path.isfile(__a ): UpperCamelCase__ : Any =torch.load(__a )['''model'''] remove_ignore_keys_(__a ) UpperCamelCase__ : Union[str, Any] =rename_fairseq_keys(__a , __a ) UpperCamelCase__ : List[Any] =os.path.join( __a , weights_name.replace('''.bin''' , F'''-{len(__a )+1:05d}-of-???.bin''' ) ) torch.save(__a , __a ) sharded_state_dicts.append(expert_state.keys() ) total_size += sum([value.numel() for key, value in expert_state.items()] ) * dtype_byte_size( expert_state[list(__a )[0]].dtype ) # Add the last block UpperCamelCase__ : Union[str, Any] =os.path.join(__a , weights_name.replace('''.bin''' , F'''-{len(__a )+1:05d}-of-???.bin''' ) ) UpperCamelCase__ : List[Any] =torch.load(switch_checkpoint_path + '''-shared.pt''' )['''model'''] remove_ignore_keys_(__a ) UpperCamelCase__ : Optional[Any] =rename_fairseq_keys(__a , __a ) UpperCamelCase__ : Union[str, Any] =shared_weights['''decoder.embed_tokens.weight'''] sharded_state_dicts.append(shared_weights.keys() ) # If we only have the shared weights (dummy model/experts saved on the same file) if len(__a ) == 1: UpperCamelCase__ : Union[str, Any] =os.path.join(__a , __a ) torch.save(__a , __a ) return {weights_name: sharded_state_dicts[0]}, None else: torch.save(__a , __a ) # Otherwise, let's build the index UpperCamelCase__ : List[Any] ={} for idx, shard in enumerate(__a ): UpperCamelCase__ : Any =weights_name.replace('''.bin''' , F'''-{idx+1:05d}-of-{len(__a ):05d}.bin''' ) UpperCamelCase__ : Union[str, Any] =os.path.join(__a , weights_name.replace('''.bin''' , F'''-{idx+1:05d}-of-???.bin''' ) ) os.rename(__a , os.path.join(__a , __a ) ) for key in shard: UpperCamelCase__ : Dict =shard_file # Add the metadata UpperCamelCase__ : List[str] ={'''total_size''': total_size} UpperCamelCase__ : int ={'''metadata''': metadata, '''weight_map''': weight_map} with open(os.path.join(__a , __a ) , '''w''' , encoding='''utf-8''' ) as f: UpperCamelCase__ : str =json.dumps(__a , indent=2 , sort_keys=__a ) + '''\n''' f.write(__a ) return metadata, index if __name__ == "__main__": _SCREAMING_SNAKE_CASE : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--nllb_moe_checkpoint_path""", default="""/home/arthur_huggingface_co/fairseq/weights/checkpoints/model_moe_54b/checkpoint_2_300000""", type=str, required=False, help="""Path to a directory containing a folder per layer. Follows the original Google format.""", ) parser.add_argument("""--dtype""", default="""float32""", type=str, required=False, help="""dtype of the saved model""") parser.add_argument( """--pytorch_dump_folder_path""", default="""/home/arthur_huggingface_co/fairseq/weights/checkpoints/hf-converted-moe-54b""", type=str, required=False, help="""Path to the output pytorch model.""", ) _SCREAMING_SNAKE_CASE : Union[str, Any] = parser.parse_args() _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : Dict = shard_on_the_fly( args.nllb_moe_checkpoint_path, args.pytorch_dump_folder_path, 1_2_8, args.dtype, ) _SCREAMING_SNAKE_CASE : List[str] = NllbMoeConfig.from_pretrained( """facebook/nllb-200-3.3B""", encoder_sparse_step=4, decoder_sparse_step=4, num_experts=1_2_8 ) config.save_pretrained(args.pytorch_dump_folder_path) _SCREAMING_SNAKE_CASE : Any = NllbMoeModel.from_pretrained(args.pytorch_dump_folder_path) print("""Done""") model.save_pretrained(args.pytorch_dump_folder_path)
365
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE : Tuple = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE : List[Any] = { """microsoft/unispeech-large-1500h-cv""": ( """https://huggingface.co/microsoft/unispeech-large-1500h-cv/resolve/main/config.json""" ), # See all UniSpeech models at https://huggingface.co/models?filter=unispeech } class __a ( snake_case__ ): """simple docstring""" SCREAMING_SNAKE_CASE_ = 'unispeech' def __init__( self : List[Any] , lowercase_ : Tuple=32 , lowercase_ : int=768 , lowercase_ : List[Any]=12 , lowercase_ : Optional[int]=12 , lowercase_ : Union[str, Any]=3072 , lowercase_ : Any="gelu" , lowercase_ : List[Any]=0.1 , lowercase_ : str=0.1 , lowercase_ : Union[str, Any]=0.1 , lowercase_ : List[str]=0.0 , lowercase_ : Tuple=0.0 , lowercase_ : Dict=0.1 , lowercase_ : List[str]=0.1 , lowercase_ : List[str]=0.0_2 , lowercase_ : int=1e-5 , lowercase_ : Dict="group" , lowercase_ : Optional[Any]="gelu" , lowercase_ : List[Any]=(512, 512, 512, 512, 512, 512, 512) , lowercase_ : Union[str, Any]=(5, 2, 2, 2, 2, 2, 2) , lowercase_ : Union[str, Any]=(10, 3, 3, 3, 3, 2, 2) , lowercase_ : Any=False , lowercase_ : Dict=128 , lowercase_ : List[str]=16 , lowercase_ : Any=False , lowercase_ : Optional[Any]=True , lowercase_ : List[str]=0.0_5 , lowercase_ : int=10 , lowercase_ : Optional[Any]=2 , lowercase_ : List[str]=0.0 , lowercase_ : List[Any]=10 , lowercase_ : Union[str, Any]=0 , lowercase_ : Dict=320 , lowercase_ : Optional[Any]=2 , lowercase_ : Union[str, Any]=0.1 , lowercase_ : Dict=100 , lowercase_ : Optional[int]=256 , lowercase_ : Dict=256 , lowercase_ : Optional[int]=0.1 , lowercase_ : str="mean" , lowercase_ : Union[str, Any]=False , lowercase_ : Any=False , lowercase_ : Union[str, Any]=256 , lowercase_ : List[str]=80 , lowercase_ : Dict=0 , lowercase_ : int=1 , lowercase_ : Union[str, Any]=2 , lowercase_ : Dict=0.5 , **lowercase_ : str , ): super().__init__(**lowercase_ , pad_token_id=lowercase_ , bos_token_id=lowercase_ , eos_token_id=lowercase_ ) UpperCamelCase__ : Dict =hidden_size UpperCamelCase__ : Optional[int] =feat_extract_norm UpperCamelCase__ : Dict =feat_extract_activation UpperCamelCase__ : Union[str, Any] =list(lowercase_ ) UpperCamelCase__ : int =list(lowercase_ ) UpperCamelCase__ : Tuple =list(lowercase_ ) UpperCamelCase__ : List[str] =conv_bias UpperCamelCase__ : Any =num_conv_pos_embeddings UpperCamelCase__ : List[Any] =num_conv_pos_embedding_groups UpperCamelCase__ : Optional[int] =len(self.conv_dim ) UpperCamelCase__ : Union[str, Any] =num_hidden_layers UpperCamelCase__ : Optional[Any] =intermediate_size UpperCamelCase__ : Any =hidden_act UpperCamelCase__ : List[Any] =num_attention_heads UpperCamelCase__ : List[Any] =hidden_dropout UpperCamelCase__ : List[Any] =attention_dropout UpperCamelCase__ : Tuple =activation_dropout UpperCamelCase__ : Any =feat_proj_dropout UpperCamelCase__ : Tuple =final_dropout UpperCamelCase__ : Tuple =layerdrop UpperCamelCase__ : int =layer_norm_eps UpperCamelCase__ : Optional[int] =initializer_range UpperCamelCase__ : Any =num_ctc_classes UpperCamelCase__ : Optional[int] =vocab_size UpperCamelCase__ : int =do_stable_layer_norm UpperCamelCase__ : Union[str, Any] =use_weighted_layer_sum UpperCamelCase__ : Tuple =classifier_proj_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==''' ''' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =''' f''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' f''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 UpperCamelCase__ : List[Any] =apply_spec_augment UpperCamelCase__ : List[Any] =mask_time_prob UpperCamelCase__ : Optional[int] =mask_time_length UpperCamelCase__ : Dict =mask_time_min_masks UpperCamelCase__ : str =mask_feature_prob UpperCamelCase__ : Union[str, Any] =mask_feature_length UpperCamelCase__ : int =mask_feature_min_masks # parameters for pretraining with codevector quantized representations UpperCamelCase__ : Optional[Any] =num_codevectors_per_group UpperCamelCase__ : Dict =num_codevector_groups UpperCamelCase__ : int =contrastive_logits_temperature UpperCamelCase__ : Tuple =feat_quantizer_dropout UpperCamelCase__ : List[str] =num_negatives UpperCamelCase__ : Dict =codevector_dim UpperCamelCase__ : Any =proj_codevector_dim UpperCamelCase__ : List[Any] =diversity_loss_weight # ctc loss UpperCamelCase__ : Tuple =ctc_loss_reduction UpperCamelCase__ : List[str] =ctc_zero_infinity # pretraining loss UpperCamelCase__ : Optional[Any] =replace_prob @property def _lowerCAmelCase ( self : List[str] ): return functools.reduce(operator.mul , self.conv_stride , 1 )
157
0
'''simple docstring''' def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" if a < 0: raise ValueError("""Input value must be a positive integer""" ) elif isinstance(lowerCAmelCase , lowerCAmelCase ): raise TypeError("""Input value must be a 'int' type""" ) return bin(lowerCAmelCase ).count("""1""" ) if __name__ == "__main__": import doctest doctest.testmod()
70
"""simple docstring""" import os def UpperCAmelCase__ ( ): """simple docstring""" A_ : Any = os.path.join(os.path.dirname(_UpperCAmelCase ) , 'num.txt' ) with open(_UpperCAmelCase ) as file_hand: return str(sum(int(_UpperCAmelCase ) for line in file_hand ) )[:10] if __name__ == "__main__": print(solution())
286
0
'''simple docstring''' # Imports import numpy as np class lowerCAmelCase_ : '''simple docstring''' def __init__( self : Tuple , _UpperCAmelCase : List[str]=None , _UpperCAmelCase : int=None , _UpperCAmelCase : Dict=None , _UpperCAmelCase : Optional[Any]=None , _UpperCAmelCase : Any=None ): """simple docstring""" self.set_matricies(red=_UpperCAmelCase , green=_UpperCAmelCase , blue=_UpperCAmelCase , red_edge=_UpperCAmelCase , nir=_UpperCAmelCase ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , _UpperCAmelCase : Optional[int]=None , _UpperCAmelCase : Union[str, Any]=None , _UpperCAmelCase : str=None , _UpperCAmelCase : int=None , _UpperCAmelCase : List[Any]=None ): """simple docstring""" if red is not None: UpperCAmelCase__ = red if green is not None: UpperCAmelCase__ = green if blue is not None: UpperCAmelCase__ = blue if red_edge is not None: UpperCAmelCase__ = red_edge if nir is not None: UpperCAmelCase__ = nir return True def SCREAMING_SNAKE_CASE__ ( self : List[Any] , _UpperCAmelCase : Tuple="" , _UpperCAmelCase : Any=None , _UpperCAmelCase : int=None , _UpperCAmelCase : Optional[int]=None , _UpperCAmelCase : Optional[int]=None , _UpperCAmelCase : Optional[int]=None ): """simple docstring""" self.set_matricies(red=_UpperCAmelCase , green=_UpperCAmelCase , blue=_UpperCAmelCase , red_edge=_UpperCAmelCase , nir=_UpperCAmelCase ) UpperCAmelCase__ = { """ARVI2""": self.arvaa, """CCCI""": self.ccci, """CVI""": self.cvi, """GLI""": self.gli, """NDVI""": self.ndvi, """BNDVI""": self.bndvi, """redEdgeNDVI""": self.red_edge_ndvi, """GNDVI""": self.gndvi, """GBNDVI""": self.gbndvi, """GRNDVI""": self.grndvi, """RBNDVI""": self.rbndvi, """PNDVI""": self.pndvi, """ATSAVI""": self.atsavi, """BWDRVI""": self.bwdrvi, """CIgreen""": self.ci_green, """CIrededge""": self.ci_rededge, """CI""": self.ci, """CTVI""": self.ctvi, """GDVI""": self.gdvi, """EVI""": self.evi, """GEMI""": self.gemi, """GOSAVI""": self.gosavi, """GSAVI""": self.gsavi, """Hue""": self.hue, """IVI""": self.ivi, """IPVI""": self.ipvi, """I""": self.i, """RVI""": self.rvi, """MRVI""": self.mrvi, """MSAVI""": self.m_savi, """NormG""": self.norm_g, """NormNIR""": self.norm_nir, """NormR""": self.norm_r, """NGRDI""": self.ngrdi, """RI""": self.ri, """S""": self.s, """IF""": self._if, """DVI""": self.dvi, """TVI""": self.tvi, """NDRE""": self.ndre, } try: return funcs[index]() except KeyError: print("""Index not in the list!""" ) return False def SCREAMING_SNAKE_CASE__ ( self : Tuple ): """simple docstring""" return -0.18 + (1.17 * ((self.nir - self.red) / (self.nir + self.red))) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ): """simple docstring""" return ((self.nir - self.redEdge) / (self.nir + self.redEdge)) / ( (self.nir - self.red) / (self.nir + self.red) ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): """simple docstring""" return self.nir * (self.red / (self.green**2)) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ): """simple docstring""" return (2 * self.green - self.red - self.blue) / ( 2 * self.green + self.red + self.blue ) def SCREAMING_SNAKE_CASE__ ( self : Any ): """simple docstring""" return (self.nir - self.red) / (self.nir + self.red) def SCREAMING_SNAKE_CASE__ ( self : Dict ): """simple docstring""" return (self.nir - self.blue) / (self.nir + self.blue) def SCREAMING_SNAKE_CASE__ ( self : int ): """simple docstring""" return (self.redEdge - self.red) / (self.redEdge + self.red) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ): """simple docstring""" return (self.nir - self.green) / (self.nir + self.green) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ): """simple docstring""" return (self.nir - (self.green + self.blue)) / ( self.nir + (self.green + self.blue) ) def SCREAMING_SNAKE_CASE__ ( self : Tuple ): """simple docstring""" return (self.nir - (self.green + self.red)) / ( self.nir + (self.green + self.red) ) def SCREAMING_SNAKE_CASE__ ( self : Dict ): """simple docstring""" return (self.nir - (self.blue + self.red)) / (self.nir + (self.blue + self.red)) def SCREAMING_SNAKE_CASE__ ( self : List[str] ): """simple docstring""" return (self.nir - (self.green + self.red + self.blue)) / ( self.nir + (self.green + self.red + self.blue) ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] , _UpperCAmelCase : Any=0.08 , _UpperCAmelCase : Any=1.22 , _UpperCAmelCase : Any=0.03 ): """simple docstring""" return a * ( (self.nir - a * self.red - b) / (a * self.nir + self.red - a * b + x * (1 + a**2)) ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): """simple docstring""" return (0.1 * self.nir - self.blue) / (0.1 * self.nir + self.blue) def SCREAMING_SNAKE_CASE__ ( self : int ): """simple docstring""" return (self.nir / self.green) - 1 def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): """simple docstring""" return (self.nir / self.redEdge) - 1 def SCREAMING_SNAKE_CASE__ ( self : int ): """simple docstring""" return (self.red - self.blue) / self.red def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): """simple docstring""" UpperCAmelCase__ = self.ndvi() return ((ndvi + 0.5) / (abs(ndvi + 0.5 ))) * (abs(ndvi + 0.5 ) ** (1 / 2)) def SCREAMING_SNAKE_CASE__ ( self : str ): """simple docstring""" return self.nir - self.green def SCREAMING_SNAKE_CASE__ ( self : Any ): """simple docstring""" return 2.5 * ( (self.nir - self.red) / (self.nir + 6 * self.red - 7.5 * self.blue + 1) ) def SCREAMING_SNAKE_CASE__ ( self : Any ): """simple docstring""" UpperCAmelCase__ = (2 * (self.nir**2 - self.red**2) + 1.5 * self.nir + 0.5 * self.red) / ( self.nir + self.red + 0.5 ) return n * (1 - 0.25 * n) - (self.red - 0.125) / (1 - self.red) def SCREAMING_SNAKE_CASE__ ( self : Tuple , _UpperCAmelCase : Dict=0.16 ): """simple docstring""" return (self.nir - self.green) / (self.nir + self.green + y) def SCREAMING_SNAKE_CASE__ ( self : List[Any] , _UpperCAmelCase : Dict=0.5 ): """simple docstring""" return ((self.nir - self.green) / (self.nir + self.green + n)) * (1 + n) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ): """simple docstring""" return np.arctan( ((2 * self.red - self.green - self.blue) / 30.5) * (self.green - self.blue) ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , _UpperCAmelCase : int=None , _UpperCAmelCase : Dict=None ): """simple docstring""" return (self.nir - b) / (a * self.red) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): """simple docstring""" return (self.nir / ((self.nir + self.red) / 2)) * (self.ndvi() + 1) def SCREAMING_SNAKE_CASE__ ( self : Dict ): """simple docstring""" return (self.red + self.green + self.blue) / 30.5 def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ): """simple docstring""" return self.nir / self.red def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): """simple docstring""" return (self.rvi() - 1) / (self.rvi() + 1) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): """simple docstring""" return ( (2 * self.nir + 1) - ((2 * self.nir + 1) ** 2 - 8 * (self.nir - self.red)) ** (1 / 2) ) / 2 def SCREAMING_SNAKE_CASE__ ( self : int ): """simple docstring""" return self.green / (self.nir + self.red + self.green) def SCREAMING_SNAKE_CASE__ ( self : Any ): """simple docstring""" return self.nir / (self.nir + self.red + self.green) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ): """simple docstring""" return self.red / (self.nir + self.red + self.green) def SCREAMING_SNAKE_CASE__ ( self : Dict ): """simple docstring""" return (self.green - self.red) / (self.green + self.red) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ): """simple docstring""" return (self.red - self.green) / (self.red + self.green) def SCREAMING_SNAKE_CASE__ ( self : Dict ): """simple docstring""" UpperCAmelCase__ = np.max([np.max(self.red ), np.max(self.green ), np.max(self.blue )] ) UpperCAmelCase__ = np.min([np.min(self.red ), np.min(self.green ), np.min(self.blue )] ) return (max_value - min_value) / max_value def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): """simple docstring""" return (2 * self.red - self.green - self.blue) / (self.green - self.blue) def SCREAMING_SNAKE_CASE__ ( self : str ): """simple docstring""" return self.nir / self.red def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): """simple docstring""" return (self.ndvi() + 0.5) ** (1 / 2) def SCREAMING_SNAKE_CASE__ ( self : int ): """simple docstring""" return (self.nir - self.redEdge) / (self.nir + self.redEdge)
61
'''simple docstring''' import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) UpperCAmelCase_ = logging.getLogger(__name__) UpperCAmelCase_ = 'Hello world! cécé herlolip' UpperCAmelCase_ = namedtuple( 'BertAbsConfig', [ 'temp_dir', 'large', 'use_bert_emb', 'finetune_bert', 'encoder', 'share_emb', 'max_pos', 'enc_layers', 'enc_hidden_size', 'enc_heads', 'enc_ff_size', 'enc_dropout', 'dec_layers', 'dec_hidden_size', 'dec_heads', 'dec_ff_size', 'dec_dropout', ], ) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Tuple ): '''simple docstring''' UpperCAmelCase__ = BertAbsConfig( temp_dir=""".""" , finetune_bert=SCREAMING_SNAKE_CASE__ , large=SCREAMING_SNAKE_CASE__ , share_emb=SCREAMING_SNAKE_CASE__ , use_bert_emb=SCREAMING_SNAKE_CASE__ , encoder="""bert""" , max_pos=512 , enc_layers=6 , enc_hidden_size=512 , enc_heads=8 , enc_ff_size=512 , enc_dropout=0.2 , dec_layers=6 , dec_hidden_size=768 , dec_heads=8 , dec_ff_size=2048 , dec_dropout=0.2 , ) UpperCAmelCase__ = torch.load(SCREAMING_SNAKE_CASE__ , lambda SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : storage ) UpperCAmelCase__ = AbsSummarizer(SCREAMING_SNAKE_CASE__ , torch.device("""cpu""" ) , SCREAMING_SNAKE_CASE__ ) original.eval() UpperCAmelCase__ = BertAbsSummarizer(SCREAMING_SNAKE_CASE__ , torch.device("""cpu""" ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info("""convert the model""" ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info("""Make sure that the models' outputs are identical""" ) UpperCAmelCase__ = BertTokenizer.from_pretrained("""bert-base-uncased""" ) # prepare the model inputs UpperCAmelCase__ = tokenizer.encode("""This is sample éàalj'-.""" ) encoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(SCREAMING_SNAKE_CASE__ )) ) UpperCAmelCase__ = torch.tensor(SCREAMING_SNAKE_CASE__ ).unsqueeze(0 ) UpperCAmelCase__ = tokenizer.encode("""This is sample 3 éàalj'-.""" ) decoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(SCREAMING_SNAKE_CASE__ )) ) UpperCAmelCase__ = torch.tensor(SCREAMING_SNAKE_CASE__ ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass UpperCAmelCase__ = encoder_input_ids UpperCAmelCase__ = decoder_input_ids UpperCAmelCase__ = UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = UpperCAmelCase__ = None UpperCAmelCase__ = UpperCAmelCase__ = None UpperCAmelCase__ = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical UpperCAmelCase__ = original(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ )[0] UpperCAmelCase__ = original.generator(SCREAMING_SNAKE_CASE__ ) UpperCAmelCase__ = new_model( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ )[0] UpperCAmelCase__ = new_model.generator(SCREAMING_SNAKE_CASE__ ) UpperCAmelCase__ = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print("""Maximum absolute difference beween weights: {:.2f}""".format(SCREAMING_SNAKE_CASE__ ) ) UpperCAmelCase__ = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print("""Maximum absolute difference beween weights: {:.2f}""".format(SCREAMING_SNAKE_CASE__ ) ) UpperCAmelCase__ = torch.allclose(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , atol=1e-3 ) if are_identical: logging.info("""all weights are equal up to 1e-3""" ) else: raise ValueError("""the weights are different. The new model is likely different from the original one.""" ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info("""saving the model's state dictionary""" ) torch.save( new_model.state_dict() , """./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin""" ) if __name__ == "__main__": UpperCAmelCase_ = argparse.ArgumentParser() parser.add_argument( '--bertabs_checkpoint_path', default=None, type=str, required=True, help='Path the official PyTorch dump.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.', ) UpperCAmelCase_ = parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
61
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) SCREAMING_SNAKE_CASE__ = {"configuration_deit": ["DEIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "DeiTConfig", "DeiTOnnxConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ = ["DeiTFeatureExtractor"] SCREAMING_SNAKE_CASE__ = ["DeiTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ = [ "DEIT_PRETRAINED_MODEL_ARCHIVE_LIST", "DeiTForImageClassification", "DeiTForImageClassificationWithTeacher", "DeiTForMaskedImageModeling", "DeiTModel", "DeiTPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ = [ "TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFDeiTForImageClassification", "TFDeiTForImageClassificationWithTeacher", "TFDeiTForMaskedImageModeling", "TFDeiTModel", "TFDeiTPreTrainedModel", ] if TYPE_CHECKING: from .configuration_deit import DEIT_PRETRAINED_CONFIG_ARCHIVE_MAP, DeiTConfig, DeiTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_deit import DeiTFeatureExtractor from .image_processing_deit import DeiTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_deit import ( DEIT_PRETRAINED_MODEL_ARCHIVE_LIST, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, DeiTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_deit import ( TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, TFDeiTPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
150
"""simple docstring""" from ...configuration_utils import PretrainedConfig class lowerCAmelCase_ ( lowerCAmelCase ): """simple docstring""" _lowerCAmelCase : Tuple = """bert-generation""" def __init__( self , lowerCAmelCase=5_03_58 , lowerCAmelCase=10_24 , lowerCAmelCase=24 , lowerCAmelCase=16 , lowerCAmelCase=40_96 , lowerCAmelCase="gelu" , lowerCAmelCase=0.1 , lowerCAmelCase=0.1 , lowerCAmelCase=5_12 , lowerCAmelCase=0.02 , lowerCAmelCase=1E-12 , lowerCAmelCase=0 , lowerCAmelCase=2 , lowerCAmelCase=1 , lowerCAmelCase="absolute" , lowerCAmelCase=True , **lowerCAmelCase , ): """simple docstring""" super().__init__(pad_token_id=lowerCAmelCase , bos_token_id=lowerCAmelCase , eos_token_id=lowerCAmelCase , **lowerCAmelCase ) snake_case = vocab_size snake_case = hidden_size snake_case = num_hidden_layers snake_case = num_attention_heads snake_case = hidden_act snake_case = intermediate_size snake_case = hidden_dropout_prob snake_case = attention_probs_dropout_prob snake_case = max_position_embeddings snake_case = initializer_range snake_case = layer_norm_eps snake_case = position_embedding_type snake_case = use_cache
150
1
"""simple docstring""" def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> int: if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): raise TypeError('Input value must be an \'int\' type' ) __lowerCamelCase : List[str] = 0 while number: position += 1 number >>= 1 return position if __name__ == "__main__": import doctest doctest.testmod()
371
import argparse import intel_extension_for_pytorch as ipex import torch from diffusers import DPMSolverMultistepScheduler, StableDiffusionPipeline a =argparse.ArgumentParser("""Stable Diffusion script with intel optimization""", add_help=False) parser.add_argument("""--dpm""", action="""store_true""", help="""Enable DPMSolver or not""") parser.add_argument("""--steps""", default=None, type=int, help="""Num inference steps""") a =parser.parse_args() a ="""cpu""" a ="""a lovely <dicoo> in red dress and hat, in the snowly and brightly night, with many brighly buildings""" a ="""path-to-your-trained-model""" a =StableDiffusionPipeline.from_pretrained(model_id) if args.dpm: a =DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) a =pipe.to(device) # to channels last a =pipe.unet.to(memory_format=torch.channels_last) a =pipe.vae.to(memory_format=torch.channels_last) a =pipe.text_encoder.to(memory_format=torch.channels_last) if pipe.requires_safety_checker: a =pipe.safety_checker.to(memory_format=torch.channels_last) # optimize with ipex a =torch.randn(2, 4, 64, 64) a =torch.rand(1) * 999 a =torch.randn(2, 77, 768) a =(sample, timestep, encoder_hidden_status) try: a =ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True, sample_input=input_example) except Exception: a =ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True) a =ipex.optimize(pipe.vae.eval(), dtype=torch.bfloataa, inplace=True) a =ipex.optimize(pipe.text_encoder.eval(), dtype=torch.bfloataa, inplace=True) if pipe.requires_safety_checker: a =ipex.optimize(pipe.safety_checker.eval(), dtype=torch.bfloataa, inplace=True) # compute a =666 a =torch.Generator(device).manual_seed(seed) a ={"""generator""": generator} if args.steps is not None: a =args.steps with torch.cpu.amp.autocast(enabled=True, dtype=torch.bfloataa): a =pipe(prompt, **generate_kwargs).images[0] # save image image.save("""generated.png""")
113
0