code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
'''simple docstring''' import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def lowerCAmelCase_ ( snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' A : int = 1.5 A : Dict = int(factor * num_class_images ) A : Tuple = ClipClient( url='''https://knn.laion.ai/knn-service''' , indice_name='''laion_400m''' , num_images=snake_case__ , aesthetic_weight=0.1 ) os.makedirs(F'{class_data_dir}/images' , exist_ok=snake_case__ ) if len(list(Path(F'{class_data_dir}/images' ).iterdir() ) ) >= num_class_images: return while True: A : Tuple = client.query(text=snake_case__ ) if len(snake_case__ ) >= factor * num_class_images or num_images > 1E4: break else: A : Dict = int(factor * num_images ) A : Optional[int] = ClipClient( url='''https://knn.laion.ai/knn-service''' , indice_name='''laion_400m''' , num_images=snake_case__ , aesthetic_weight=0.1 , ) A : Dict = 0 A : Optional[int] = 0 A : Tuple = tqdm(desc='''downloading real regularization images''' , total=snake_case__ ) with open(F'{class_data_dir}/caption.txt' , '''w''' ) as fa, open(F'{class_data_dir}/urls.txt' , '''w''' ) as fa, open( F'{class_data_dir}/images.txt' , '''w''' ) as fa: while total < num_class_images: A : Optional[Any] = class_images[count] count += 1 try: A : Dict = requests.get(images['''url'''] ) if img.status_code == 200: A : Tuple = Image.open(BytesIO(img.content ) ) with open(F'{class_data_dir}/images/{total}.jpg' , '''wb''' ) as f: f.write(img.content ) fa.write(images['''caption'''] + '''\n''' ) fa.write(images['''url'''] + '''\n''' ) fa.write(F'{class_data_dir}/images/{total}.jpg' + '''\n''' ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def lowerCAmelCase_ ( ): '''simple docstring''' A : int = argparse.ArgumentParser('''''' , add_help=snake_case__ ) parser.add_argument('''--class_prompt''' , help='''text prompt to retrieve images''' , required=snake_case__ , type=snake_case__ ) parser.add_argument('''--class_data_dir''' , help='''path to save images''' , required=snake_case__ , type=snake_case__ ) parser.add_argument('''--num_class_images''' , help='''number of images to download''' , default=200 , type=snake_case__ ) return parser.parse_args() if __name__ == "__main__": lowercase : Optional[int] = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
3
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase : Union[str, Any] = logging.get_logger(__name__) lowercase : str = { 'bert-base-uncased': 'https://huggingface.co/bert-base-uncased/resolve/main/config.json', 'bert-large-uncased': 'https://huggingface.co/bert-large-uncased/resolve/main/config.json', 'bert-base-cased': 'https://huggingface.co/bert-base-cased/resolve/main/config.json', 'bert-large-cased': 'https://huggingface.co/bert-large-cased/resolve/main/config.json', 'bert-base-multilingual-uncased': 'https://huggingface.co/bert-base-multilingual-uncased/resolve/main/config.json', 'bert-base-multilingual-cased': 'https://huggingface.co/bert-base-multilingual-cased/resolve/main/config.json', 'bert-base-chinese': 'https://huggingface.co/bert-base-chinese/resolve/main/config.json', 'bert-base-german-cased': 'https://huggingface.co/bert-base-german-cased/resolve/main/config.json', 'bert-large-uncased-whole-word-masking': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/config.json' ), 'bert-large-cased-whole-word-masking': ( 'https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/config.json' ), 'bert-large-uncased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/config.json' ), 'bert-large-cased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/config.json' ), 'bert-base-cased-finetuned-mrpc': 'https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/config.json', 'bert-base-german-dbmdz-cased': 'https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/config.json', 'bert-base-german-dbmdz-uncased': 'https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/config.json', 'cl-tohoku/bert-base-japanese': 'https://huggingface.co/cl-tohoku/bert-base-japanese/resolve/main/config.json', 'cl-tohoku/bert-base-japanese-whole-word-masking': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-whole-word-masking/resolve/main/config.json' ), 'cl-tohoku/bert-base-japanese-char': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-char/resolve/main/config.json' ), 'cl-tohoku/bert-base-japanese-char-whole-word-masking': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-char-whole-word-masking/resolve/main/config.json' ), 'TurkuNLP/bert-base-finnish-cased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/config.json' ), 'TurkuNLP/bert-base-finnish-uncased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/config.json' ), 'wietsedv/bert-base-dutch-cased': 'https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/config.json', # See all BERT models at https://huggingface.co/models?filter=bert } class A ( __snake_case ): __magic_name__ = '''bert''' def __init__( self , SCREAMING_SNAKE_CASE=30522 , SCREAMING_SNAKE_CASE=768 , SCREAMING_SNAKE_CASE=12 , SCREAMING_SNAKE_CASE=12 , SCREAMING_SNAKE_CASE=3072 , SCREAMING_SNAKE_CASE="gelu" , SCREAMING_SNAKE_CASE=0.1 , SCREAMING_SNAKE_CASE=0.1 , SCREAMING_SNAKE_CASE=512 , SCREAMING_SNAKE_CASE=2 , SCREAMING_SNAKE_CASE=0.02 , SCREAMING_SNAKE_CASE=1e-12 , SCREAMING_SNAKE_CASE=0 , SCREAMING_SNAKE_CASE="absolute" , SCREAMING_SNAKE_CASE=True , SCREAMING_SNAKE_CASE=None , **SCREAMING_SNAKE_CASE , ) -> Optional[int]: """simple docstring""" super().__init__(pad_token_id=SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) A : Optional[int] = vocab_size A : Optional[Any] = hidden_size A : List[Any] = num_hidden_layers A : List[str] = num_attention_heads A : Dict = hidden_act A : Optional[Any] = intermediate_size A : List[Any] = hidden_dropout_prob A : List[Any] = attention_probs_dropout_prob A : Optional[Any] = max_position_embeddings A : List[str] = type_vocab_size A : Dict = initializer_range A : str = layer_norm_eps A : int = position_embedding_type A : Dict = use_cache A : str = classifier_dropout class A ( __snake_case ): @property def __lowerCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "multiple-choice": A : Optional[Any] = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: A : Optional[int] = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ('''token_type_ids''', dynamic_axis), ] )
3
1
import torch from diffusers import KDPMaDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class SCREAMING_SNAKE_CASE ( lowerCAmelCase ): '''simple docstring''' UpperCamelCase_ : Any = (KDPMaDiscreteScheduler,) UpperCamelCase_ : Any = 1_0 def _A ( self : List[str] , **UpperCAmelCase_ : List[Any] ): SCREAMING_SNAKE_CASE : Any = { "num_train_timesteps": 1100, "beta_start": 0.0_001, "beta_end": 0.02, "beta_schedule": "linear", } config.update(**UpperCAmelCase_ ) return config def _A ( self : List[str] ): for timesteps in [10, 50, 100, 1000]: self.check_over_configs(num_train_timesteps=UpperCAmelCase_ ) def _A ( self : Optional[Any] ): for beta_start, beta_end in zip([0.00_001, 0.0_001, 0.001] , [0.0_002, 0.002, 0.02] ): self.check_over_configs(beta_start=UpperCAmelCase_ , beta_end=UpperCAmelCase_ ) def _A ( self : Tuple ): for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=UpperCAmelCase_ ) def _A ( self : str ): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=UpperCAmelCase_ ) def _A ( self : Optional[int] ): SCREAMING_SNAKE_CASE : List[Any] = self.scheduler_classes[0] SCREAMING_SNAKE_CASE : List[str] = self.get_scheduler_config(prediction_type="v_prediction" ) SCREAMING_SNAKE_CASE : Optional[Any] = scheduler_class(**UpperCAmelCase_ ) scheduler.set_timesteps(self.num_inference_steps ) SCREAMING_SNAKE_CASE : Any = self.dummy_model() SCREAMING_SNAKE_CASE : Tuple = self.dummy_sample_deter * scheduler.init_noise_sigma SCREAMING_SNAKE_CASE : Tuple = sample.to(UpperCAmelCase_ ) for i, t in enumerate(scheduler.timesteps ): SCREAMING_SNAKE_CASE : Optional[int] = scheduler.scale_model_input(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : int = model(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : List[Any] = scheduler.step(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : Tuple = output.prev_sample SCREAMING_SNAKE_CASE : str = torch.sum(torch.abs(UpperCAmelCase_ ) ) SCREAMING_SNAKE_CASE : Optional[int] = torch.mean(torch.abs(UpperCAmelCase_ ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 4.6_934E-07 ) < 1E-2 assert abs(result_mean.item() - 6.1_112E-10 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 4.693_428_650_170_972E-07 ) < 1E-2 assert abs(result_mean.item() - 0.0_002 ) < 1E-3 def _A ( self : Any ): if torch_device == "mps": return SCREAMING_SNAKE_CASE : List[str] = self.scheduler_classes[0] SCREAMING_SNAKE_CASE : List[Any] = self.get_scheduler_config() SCREAMING_SNAKE_CASE : Optional[Any] = scheduler_class(**UpperCAmelCase_ ) scheduler.set_timesteps(self.num_inference_steps ) SCREAMING_SNAKE_CASE : int = self.dummy_model() SCREAMING_SNAKE_CASE : Optional[Any] = self.dummy_sample_deter * scheduler.init_noise_sigma SCREAMING_SNAKE_CASE : List[Any] = sample.to(UpperCAmelCase_ ) for i, t in enumerate(scheduler.timesteps ): SCREAMING_SNAKE_CASE : Tuple = scheduler.scale_model_input(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : Optional[Any] = model(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : int = scheduler.step(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : Optional[int] = output.prev_sample SCREAMING_SNAKE_CASE : Optional[Any] = torch.sum(torch.abs(UpperCAmelCase_ ) ) SCREAMING_SNAKE_CASE : int = torch.mean(torch.abs(UpperCAmelCase_ ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 20.4_125 ) < 1E-2 assert abs(result_mean.item() - 0.0_266 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 20.4_125 ) < 1E-2 assert abs(result_mean.item() - 0.0_266 ) < 1E-3 def _A ( self : Dict ): if torch_device == "mps": return SCREAMING_SNAKE_CASE : Dict = self.scheduler_classes[0] SCREAMING_SNAKE_CASE : List[Any] = self.get_scheduler_config() SCREAMING_SNAKE_CASE : Any = scheduler_class(**UpperCAmelCase_ ) scheduler.set_timesteps(self.num_inference_steps , device=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : Optional[Any] = self.dummy_model() SCREAMING_SNAKE_CASE : List[str] = self.dummy_sample_deter.to(UpperCAmelCase_ ) * scheduler.init_noise_sigma for t in scheduler.timesteps: SCREAMING_SNAKE_CASE : Union[str, Any] = scheduler.scale_model_input(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : Tuple = model(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : Optional[int] = scheduler.step(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE : Optional[int] = output.prev_sample SCREAMING_SNAKE_CASE : Any = torch.sum(torch.abs(UpperCAmelCase_ ) ) SCREAMING_SNAKE_CASE : Optional[Any] = torch.mean(torch.abs(UpperCAmelCase_ ) ) if str(UpperCAmelCase_ ).startswith("cpu" ): # The following sum varies between 148 and 156 on mps. Why? assert abs(result_sum.item() - 20.4_125 ) < 1E-2 assert abs(result_mean.item() - 0.0_266 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 20.4_125 ) < 1E-2 assert abs(result_mean.item() - 0.0_266 ) < 1E-3
319
# limitations under the License. # NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from .pipelines import DiffusionPipeline, ImagePipelineOutput # noqa: F401 from .utils import deprecate deprecate( """pipelines_utils""", """0.22.0""", """Importing `DiffusionPipeline` or `ImagePipelineOutput` from diffusers.pipeline_utils is deprecated. Please import from diffusers.pipelines.pipeline_utils instead.""", standard_warn=False, stacklevel=3, )
319
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _lowerCamelCase : Optional[Any] = { 'configuration_groupvit': [ 'GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GroupViTConfig', 'GroupViTOnnxConfig', 'GroupViTTextConfig', 'GroupViTVisionConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Optional[Any] = [ 'GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'GroupViTModel', 'GroupViTPreTrainedModel', 'GroupViTTextModel', 'GroupViTVisionModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Any = [ 'TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFGroupViTModel', 'TFGroupViTPreTrainedModel', 'TFGroupViTTextModel', 'TFGroupViTVisionModel', ] if TYPE_CHECKING: from .configuration_groupvit import ( GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GroupViTConfig, GroupViTOnnxConfig, GroupViTTextConfig, GroupViTVisionConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_groupvit import ( GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, GroupViTModel, GroupViTPreTrainedModel, GroupViTTextModel, GroupViTVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_groupvit import ( TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFGroupViTModel, TFGroupViTPreTrainedModel, TFGroupViTTextModel, TFGroupViTVisionModel, ) else: import sys _lowerCamelCase : Optional[int] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
258
'''simple docstring''' import itertools from dataclasses import dataclass from typing import Optional import pandas as pd import pyarrow as pa import datasets from datasets.table import table_cast @dataclass class __UpperCAmelCase ( datasets.BuilderConfig ): '''simple docstring''' __lowerCAmelCase = None class __UpperCAmelCase ( datasets.ArrowBasedBuilder ): '''simple docstring''' __lowerCAmelCase = PandasConfig def A (self : Tuple ): return datasets.DatasetInfo(features=self.config.features ) def A (self : Optional[int] , _lowerCAmelCase : List[Any] ): if not self.config.data_files: raise ValueError(F"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) A = dl_manager.download_and_extract(self.config.data_files ) if isinstance(_lowerCAmelCase , (str, list, tuple) ): A = data_files if isinstance(_lowerCAmelCase , _lowerCAmelCase ): A = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive A = [dl_manager.iter_files(_lowerCAmelCase ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"""files""": files} )] A = [] for split_name, files in data_files.items(): if isinstance(_lowerCAmelCase , _lowerCAmelCase ): A = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive A = [dl_manager.iter_files(_lowerCAmelCase ) for file in files] splits.append(datasets.SplitGenerator(name=_lowerCAmelCase , gen_kwargs={"""files""": files} ) ) return splits def A (self : Dict , _lowerCAmelCase : pa.Table ): if self.config.features is not None: # more expensive cast to support nested features with keys in a different order # allows str <-> int/float or str to Audio for example A = table_cast(_lowerCAmelCase , self.config.features.arrow_schema ) return pa_table def A (self : List[Any] , _lowerCAmelCase : Optional[Any] ): for i, file in enumerate(itertools.chain.from_iterable(_lowerCAmelCase ) ): with open(_lowerCAmelCase , """rb""" ) as f: A = pa.Table.from_pandas(pd.read_pickle(_lowerCAmelCase ) ) yield i, self._cast_table(_lowerCAmelCase )
258
1
"""simple docstring""" import re import time from typing import Optional import IPython.display as disp from ..trainer_callback import TrainerCallback from ..trainer_utils import IntervalStrategy, has_length def a__ ( SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' lowerCAmelCase : str = int(SCREAMING_SNAKE_CASE ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase : Optional[int] = t // 3_6_0_0, (t // 6_0) % 6_0, t % 6_0 return f"""{h}:{m:02d}:{s:02d}""" if h != 0 else f"""{m:02d}:{s:02d}""" def a__ ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : Optional[int] , SCREAMING_SNAKE_CASE : Optional[int] , SCREAMING_SNAKE_CASE : List[str] , SCREAMING_SNAKE_CASE : Union[str, Any]=3_0_0 ): '''simple docstring''' return f""" <div> {prefix} <progress value='{value}' max='{total}' style='width:{width}px; height:20px; vertical-align: middle;'></progress> {label} </div> """ def a__ ( SCREAMING_SNAKE_CASE : Dict ): '''simple docstring''' lowerCAmelCase : Tuple = "<table border=\"1\" class=\"dataframe\">\n" html_code += """ <thead>\n <tr style="text-align: left;">\n""" for i in items[0]: html_code += f""" <th>{i}</th>\n""" html_code += " </tr>\n </thead>\n <tbody>\n" for line in items[1:]: html_code += " <tr>\n" for elt in line: lowerCAmelCase : Any = f"""{elt:.6f}""" if isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) else str(SCREAMING_SNAKE_CASE ) html_code += f""" <td>{elt}</td>\n""" html_code += " </tr>\n" html_code += " </tbody>\n</table><p>" return html_code class SCREAMING_SNAKE_CASE__ : """simple docstring""" a : int =5 a : List[Any] =0.2 def __init__( self , snake_case__ , snake_case__ = None , snake_case__ = True , snake_case__ = None , snake_case__ = 300 , ): """simple docstring""" lowerCAmelCase : Union[str, Any] = total lowerCAmelCase : str = "" if prefix is None else prefix lowerCAmelCase : Optional[Any] = leave lowerCAmelCase : List[str] = parent lowerCAmelCase : Optional[int] = width lowerCAmelCase : Tuple = None lowerCAmelCase : List[str] = None lowerCAmelCase : Optional[int] = None def lowercase__ ( self , snake_case__ , snake_case__ = False , snake_case__ = None ): """simple docstring""" lowerCAmelCase : Optional[int] = value if comment is not None: lowerCAmelCase : str = comment if self.last_value is None: lowerCAmelCase : int = time.time() lowerCAmelCase : Any = value lowerCAmelCase : Optional[int] = None lowerCAmelCase : Optional[int] = self.warmup lowerCAmelCase : Union[str, Any] = 1 self.update_bar(snake_case__ ) elif value <= self.last_value and not force_update: return elif force_update or self.first_calls > 0 or value >= min(self.last_value + self.wait_for , self.total ): if self.first_calls > 0: self.first_calls -= 1 lowerCAmelCase : Optional[Any] = time.time() lowerCAmelCase : int = current_time - self.start_time # We could have value = self.start_value if the update is called twixe with the same start value. if value > self.start_value: lowerCAmelCase : Optional[int] = self.elapsed_time / (value - self.start_value) else: lowerCAmelCase : Optional[int] = None if value >= self.total: lowerCAmelCase : Any = self.total lowerCAmelCase : List[str] = None if not self.leave: self.close() elif self.average_time_per_item is not None: lowerCAmelCase : Tuple = self.average_time_per_item * (self.total - value) self.update_bar(snake_case__ ) lowerCAmelCase : str = value lowerCAmelCase : int = current_time if self.average_time_per_item is None: lowerCAmelCase : Union[str, Any] = 1 else: lowerCAmelCase : Union[str, Any] = max(int(self.update_every / self.average_time_per_item ) , 1 ) def lowercase__ ( self , snake_case__ , snake_case__=None ): """simple docstring""" lowerCAmelCase : Optional[int] = " " * (len(str(self.total ) ) - len(str(snake_case__ ) )) + str(snake_case__ ) if self.elapsed_time is None: lowerCAmelCase : Dict = f"""[{spaced_value}/{self.total} : < :""" elif self.predicted_remaining is None: lowerCAmelCase : str = f"""[{spaced_value}/{self.total} {format_time(self.elapsed_time )}""" else: lowerCAmelCase : Tuple = ( f"""[{spaced_value}/{self.total} {format_time(self.elapsed_time )} <""" f""" {format_time(self.predicted_remaining )}""" ) self.label += f""", {1/self.average_time_per_item:.2f} it/s""" self.label += "]" if self.comment is None or len(self.comment ) == 0 else f""", {self.comment}]""" self.display() def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Optional[int] = html_progress_bar(self.value , self.total , self.prefix , self.label , self.width ) if self.parent is not None: # If this is a child bar, the parent will take care of the display. self.parent.display() return if self.output is None: lowerCAmelCase : int = disp.display(disp.HTML(self.html_code ) , display_id=snake_case__ ) else: self.output.update(disp.HTML(self.html_code ) ) def lowercase__ ( self ): """simple docstring""" if self.parent is None and self.output is not None: self.output.update(disp.HTML("" ) ) class SCREAMING_SNAKE_CASE__ ( lowercase ): """simple docstring""" def __init__( self , snake_case__ , snake_case__=None ): """simple docstring""" super().__init__(snake_case__ ) lowerCAmelCase : Tuple = None if column_names is None else [column_names] lowerCAmelCase : Dict = None def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Optional[int] = html_progress_bar(self.value , self.total , self.prefix , self.label , self.width ) if self.inner_table is not None: self.html_code += text_to_html_table(self.inner_table ) if self.child_bar is not None: self.html_code += self.child_bar.html_code if self.output is None: lowerCAmelCase : Union[str, Any] = disp.display(disp.HTML(self.html_code ) , display_id=snake_case__ ) else: self.output.update(disp.HTML(self.html_code ) ) def lowercase__ ( self , snake_case__ ): """simple docstring""" if self.inner_table is None: lowerCAmelCase : List[str] = [list(values.keys() ), list(values.values() )] else: lowerCAmelCase : str = self.inner_table[0] if len(self.inner_table ) == 1: # We give a chance to update the column names at the first iteration for key in values.keys(): if key not in columns: columns.append(snake_case__ ) lowerCAmelCase : List[Any] = columns self.inner_table.append([values[c] for c in columns] ) def lowercase__ ( self , snake_case__ , snake_case__=None , snake_case__=300 ): """simple docstring""" lowerCAmelCase : Tuple = NotebookProgressBar(snake_case__ , prefix=snake_case__ , parent=self , width=snake_case__ ) return self.child_bar def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : str = None self.display() class SCREAMING_SNAKE_CASE__ ( lowercase ): """simple docstring""" def __init__( self ): """simple docstring""" lowerCAmelCase : Dict = None lowerCAmelCase : int = None lowerCAmelCase : Dict = False def lowercase__ ( self , snake_case__ , snake_case__ , snake_case__ , **snake_case__ ): """simple docstring""" lowerCAmelCase : List[str] = "Epoch" if args.evaluation_strategy == IntervalStrategy.EPOCH else "Step" lowerCAmelCase : str = 0 lowerCAmelCase : Optional[Any] = 0 lowerCAmelCase : Union[str, Any] = [self.first_column] + ["Training Loss"] if args.evaluation_strategy != IntervalStrategy.NO: column_names.append("Validation Loss" ) lowerCAmelCase : Optional[Any] = NotebookTrainingTracker(state.max_steps , snake_case__ ) def lowercase__ ( self , snake_case__ , snake_case__ , snake_case__ , **snake_case__ ): """simple docstring""" lowerCAmelCase : List[Any] = int(state.epoch ) if int(state.epoch ) == state.epoch else f"""{state.epoch:.2f}""" self.training_tracker.update( state.global_step + 1 , comment=f"""Epoch {epoch}/{state.num_train_epochs}""" , force_update=self._force_next_update , ) lowerCAmelCase : str = False def lowercase__ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__=None , **snake_case__ ): """simple docstring""" if not has_length(snake_case__ ): return if self.prediction_bar is None: if self.training_tracker is not None: lowerCAmelCase : Optional[Any] = self.training_tracker.add_child(len(snake_case__ ) ) else: lowerCAmelCase : Union[str, Any] = NotebookProgressBar(len(snake_case__ ) ) self.prediction_bar.update(1 ) else: self.prediction_bar.update(self.prediction_bar.value + 1 ) def lowercase__ ( self , snake_case__ , snake_case__ , snake_case__ , **snake_case__ ): """simple docstring""" if self.prediction_bar is not None: self.prediction_bar.close() lowerCAmelCase : List[str] = None def lowercase__ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__=None , **snake_case__ ): """simple docstring""" if args.evaluation_strategy == IntervalStrategy.NO and "loss" in logs: lowerCAmelCase : str = {"Training Loss": logs["loss"]} # First column is necessarily Step sine we're not in epoch eval strategy lowerCAmelCase : Optional[Any] = state.global_step self.training_tracker.write_line(snake_case__ ) def lowercase__ ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__=None , **snake_case__ ): """simple docstring""" if self.training_tracker is not None: lowerCAmelCase : Union[str, Any] = {"Training Loss": "No log", "Validation Loss": "No log"} for log in reversed(state.log_history ): if "loss" in log: lowerCAmelCase : Optional[int] = log["loss"] break if self.first_column == "Epoch": lowerCAmelCase : Dict = int(state.epoch ) else: lowerCAmelCase : Any = state.global_step lowerCAmelCase : int = "eval" for k in metrics: if k.endswith("_loss" ): lowerCAmelCase : int = re.sub(r"\_loss$" , "" , snake_case__ ) lowerCAmelCase : int = metrics.pop("total_flos" , snake_case__ ) lowerCAmelCase : int = metrics.pop("epoch" , snake_case__ ) lowerCAmelCase : Tuple = metrics.pop(f"""{metric_key_prefix}_runtime""" , snake_case__ ) lowerCAmelCase : Any = metrics.pop(f"""{metric_key_prefix}_samples_per_second""" , snake_case__ ) lowerCAmelCase : Any = metrics.pop(f"""{metric_key_prefix}_steps_per_second""" , snake_case__ ) lowerCAmelCase : Tuple = metrics.pop(f"""{metric_key_prefix}_jit_compilation_time""" , snake_case__ ) for k, v in metrics.items(): if k == f"""{metric_key_prefix}_loss""": lowerCAmelCase : str = v else: lowerCAmelCase : Dict = k.split("_" ) lowerCAmelCase : Union[str, Any] = " ".join([part.capitalize() for part in splits[1:]] ) lowerCAmelCase : Dict = v self.training_tracker.write_line(snake_case__ ) self.training_tracker.remove_child() lowerCAmelCase : List[Any] = None # Evaluation takes a long time so we should force the next update. lowerCAmelCase : int = True def lowercase__ ( self , snake_case__ , snake_case__ , snake_case__ , **snake_case__ ): """simple docstring""" self.training_tracker.update( state.global_step , comment=f"""Epoch {int(state.epoch )}/{state.num_train_epochs}""" , force_update=snake_case__ ) lowerCAmelCase : Union[str, Any] = None
133
"""simple docstring""" import inspect import unittest from typing import List import numpy as np from transformers import EfficientFormerConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, ) from transformers.models.efficientformer.modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_vision_available(): from PIL import Image from transformers import EfficientFormerImageProcessor class SCREAMING_SNAKE_CASE__ : """simple docstring""" def __init__( self , snake_case__ , snake_case__ = 13 , snake_case__ = 64 , snake_case__ = 2 , snake_case__ = 3 , snake_case__ = 3 , snake_case__ = True , snake_case__ = True , snake_case__ = 128 , snake_case__=[16, 32, 64, 128] , snake_case__ = 7 , snake_case__ = 4 , snake_case__ = 37 , snake_case__ = "gelu" , snake_case__ = 0.1 , snake_case__ = 0.1 , snake_case__ = 10 , snake_case__ = 0.02 , snake_case__ = 2 , snake_case__ = 1 , snake_case__ = 128 , snake_case__ = [2, 2, 2, 2] , snake_case__ = 2 , snake_case__ = 2 , ): """simple docstring""" lowerCAmelCase : Optional[Any] = parent lowerCAmelCase : str = batch_size lowerCAmelCase : int = image_size lowerCAmelCase : int = patch_size lowerCAmelCase : Union[str, Any] = num_channels lowerCAmelCase : int = is_training lowerCAmelCase : Tuple = use_labels lowerCAmelCase : List[Any] = hidden_size lowerCAmelCase : Dict = num_hidden_layers lowerCAmelCase : Tuple = num_attention_heads lowerCAmelCase : List[Any] = intermediate_size lowerCAmelCase : int = hidden_act lowerCAmelCase : str = hidden_dropout_prob lowerCAmelCase : Tuple = attention_probs_dropout_prob lowerCAmelCase : Optional[int] = type_sequence_label_size lowerCAmelCase : List[str] = initializer_range lowerCAmelCase : List[str] = encoder_stride lowerCAmelCase : Union[str, Any] = num_attention_outputs lowerCAmelCase : Any = embed_dim lowerCAmelCase : Tuple = embed_dim + 1 lowerCAmelCase : str = resolution lowerCAmelCase : Optional[Any] = depths lowerCAmelCase : Any = hidden_sizes lowerCAmelCase : List[str] = dim lowerCAmelCase : str = mlp_expansion_ratio def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase : Tuple = None if self.use_labels: lowerCAmelCase : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase : Tuple = self.get_config() return config, pixel_values, labels def lowercase__ ( self ): """simple docstring""" return EfficientFormerConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=snake_case__ , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , resolution=self.resolution , depths=self.depths , hidden_sizes=self.hidden_sizes , dim=self.dim , mlp_expansion_ratio=self.mlp_expansion_ratio , ) def lowercase__ ( self , snake_case__ , snake_case__ , snake_case__ ): """simple docstring""" lowerCAmelCase : Tuple = TFEfficientFormerModel(config=snake_case__ ) lowerCAmelCase : Optional[int] = model(snake_case__ , training=snake_case__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase__ ( self , snake_case__ , snake_case__ , snake_case__ ): """simple docstring""" lowerCAmelCase : Tuple = self.type_sequence_label_size lowerCAmelCase : Dict = TFEfficientFormerForImageClassification(snake_case__ ) lowerCAmelCase : Tuple = model(snake_case__ , labels=snake_case__ , training=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowerCAmelCase : str = 1 lowerCAmelCase : Any = TFEfficientFormerForImageClassification(snake_case__ ) lowerCAmelCase : List[str] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase : str = model(snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Optional[Any] = self.prepare_config_and_inputs() lowerCAmelCase , lowerCAmelCase , lowerCAmelCase : int = config_and_inputs lowerCAmelCase : List[str] = {"pixel_values": pixel_values} return config, inputs_dict @require_tf class SCREAMING_SNAKE_CASE__ ( lowercase , lowercase , unittest.TestCase ): """simple docstring""" a : Optional[int] =( ( TFEfficientFormerModel, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerForImageClassification, ) if is_tf_available() else () ) a : Union[str, Any] =( { "feature-extraction": TFEfficientFormerModel, "image-classification": ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, ), } if is_tf_available() else {} ) a : int =False a : Optional[Any] =False a : List[Any] =False a : str =False a : List[Any] =False def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Any = TFEfficientFormerModelTester(self ) lowerCAmelCase : Dict = ConfigTester( self , config_class=snake_case__ , has_text_modality=snake_case__ , hidden_size=37 ) def lowercase__ ( self ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="EfficientFormer does not use inputs_embeds" ) def lowercase__ ( self ): """simple docstring""" pass @unittest.skip(reason="EfficientFormer does not support input and output embeddings" ) def lowercase__ ( self ): """simple docstring""" pass def lowercase__ ( self ): """simple docstring""" lowerCAmelCase , lowerCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase : Tuple = model_class(snake_case__ ) lowerCAmelCase : str = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase : Optional[int] = [*signature.parameters.keys()] lowerCAmelCase : str = ["pixel_values"] self.assertListEqual(arg_names[:1] , snake_case__ ) def lowercase__ ( self ): """simple docstring""" def check_hidden_states_output(snake_case__ , snake_case__ , snake_case__ ): lowerCAmelCase : List[Any] = model_class(snake_case__ ) lowerCAmelCase : List[str] = model(**self._prepare_for_class(snake_case__ , snake_case__ ) , training=snake_case__ ) lowerCAmelCase : Dict = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowerCAmelCase : List[Any] = getattr( self.model_tester , "expected_num_hidden_layers" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(snake_case__ ) , snake_case__ ) if hasattr(self.model_tester , "encoder_seq_length" ): lowerCAmelCase : Union[str, Any] = self.model_tester.encoder_seq_length if hasattr(self.model_tester , "chunk_length" ) and self.model_tester.chunk_length > 1: lowerCAmelCase : Tuple = seq_length * self.model_tester.chunk_length else: lowerCAmelCase : List[str] = self.model_tester.seq_length self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) if config.is_encoder_decoder: lowerCAmelCase : Tuple = outputs.decoder_hidden_states self.asseretIsInstance(snake_case__ , (list, tuple) ) self.assertEqual(len(snake_case__ ) , snake_case__ ) lowerCAmelCase : int = getattr(self.model_tester , "seq_length" , snake_case__ ) lowerCAmelCase : Tuple = getattr(self.model_tester , "decoder_seq_length" , snake_case__ ) self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [decoder_seq_length, self.model_tester.hidden_size] , ) lowerCAmelCase , lowerCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase : str = True check_hidden_states_output(snake_case__ , snake_case__ , snake_case__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase : Dict = True check_hidden_states_output(snake_case__ , snake_case__ , snake_case__ ) def lowercase__ ( self , snake_case__ , snake_case__ , snake_case__=False ): """simple docstring""" lowerCAmelCase : Optional[int] = super()._prepare_for_class(snake_case__ , snake_case__ , return_labels=snake_case__ ) if return_labels: if model_class.__name__ == "TFEfficientFormerForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__ ) @unittest.skip(reason="EfficientFormer does not implement masked image modeling yet" ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*snake_case__ ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case__ ) @slow def lowercase__ ( self ): """simple docstring""" for model_name in TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase : Optional[Any] = TFEfficientFormerModel.from_pretrained(snake_case__ ) self.assertIsNotNone(snake_case__ ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase , lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase : int = True lowerCAmelCase : Union[str, Any] = getattr(self.model_tester , "seq_length" , snake_case__ ) lowerCAmelCase : Dict = getattr(self.model_tester , "encoder_seq_length" , snake_case__ ) lowerCAmelCase : Union[str, Any] = getattr(self.model_tester , "key_length" , snake_case__ ) lowerCAmelCase : List[str] = getattr(self.model_tester , "chunk_length" , snake_case__ ) if chunk_length is not None and hasattr(self.model_tester , "num_hashes" ): lowerCAmelCase : Dict = encoder_seq_length * self.model_tester.num_hashes for model_class in self.all_model_classes: lowerCAmelCase : int = True lowerCAmelCase : int = False lowerCAmelCase : Dict = True lowerCAmelCase : List[Any] = model_class(snake_case__ ) lowerCAmelCase : Tuple = model(**self._prepare_for_class(snake_case__ , snake_case__ ) , training=snake_case__ ) lowerCAmelCase : str = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(snake_case__ ) , self.model_tester.num_attention_outputs ) # check that output_attentions also work using config del inputs_dict["output_attentions"] lowerCAmelCase : int = True lowerCAmelCase : Dict = model_class(snake_case__ ) lowerCAmelCase : int = model(**self._prepare_for_class(snake_case__ , snake_case__ ) , training=snake_case__ ) lowerCAmelCase : int = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(snake_case__ ) , self.model_tester.num_attention_outputs ) if chunk_length is not None: self.assertListEqual( list(attentions[0].shape[-4:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, chunk_length, encoder_key_length] , ) else: self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, encoder_key_length] , ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase , lowerCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # Prepare our model lowerCAmelCase : List[str] = model_class(snake_case__ ) # These are maximally general inputs for the model, with multiple None dimensions # Hopefully this will catch any conditionals that fail for flexible shapes lowerCAmelCase : Optional[Any] = { key: tf.keras.Input(shape=val.shape[1:] , dtype=val.dtype , name=snake_case__ ) for key, val in model.input_signature.items() if key in model.dummy_inputs } lowerCAmelCase : Optional[Any] = model(snake_case__ ) self.assertTrue(outputs_dict is not None ) def a__ ( ): '''simple docstring''' lowerCAmelCase : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_tf @require_vision class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): """simple docstring""" @cached_property def lowercase__ ( self ): """simple docstring""" return ( EfficientFormerImageProcessor.from_pretrained("snap-research/efficientformer-l1-300" ) if is_vision_available() else None ) @slow def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Optional[Any] = TFEfficientFormerForImageClassification.from_pretrained("snap-research/efficientformer-l1-300" ) lowerCAmelCase : Tuple = self.default_image_processor lowerCAmelCase : Tuple = prepare_img() lowerCAmelCase : List[str] = image_processor(images=snake_case__ , return_tensors="tf" ) # forward pass lowerCAmelCase : Optional[int] = model(**snake_case__ , training=snake_case__ ) # verify the logits lowerCAmelCase : Union[str, Any] = tf.TensorShape((1, 1_000) ) self.assertEqual(outputs.logits.shape , snake_case__ ) lowerCAmelCase : Tuple = tf.constant([-0.0555, 0.4825, -0.0852] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , snake_case__ , atol=1e-4 ) ) @slow def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : List[Any] = TFEfficientFormerForImageClassificationWithTeacher.from_pretrained( "snap-research/efficientformer-l1-300" ) lowerCAmelCase : Optional[int] = self.default_image_processor lowerCAmelCase : Optional[Any] = prepare_img() lowerCAmelCase : Tuple = image_processor(images=snake_case__ , return_tensors="tf" ) # forward pass lowerCAmelCase : Dict = model(**snake_case__ , training=snake_case__ ) # verify the logits lowerCAmelCase : Optional[int] = tf.TensorShape((1, 1_000) ) self.assertEqual(outputs.logits.shape , snake_case__ ) lowerCAmelCase : Any = tf.constant([-0.1312, 0.4353, -1.0499] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , snake_case__ , atol=1e-4 ) )
133
1
import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py lowercase__ :str = "." # Internal TensorFlow ops that can be safely ignored (mostly specific to a saved model) lowercase__ :List[Any] = [ "Assert", "AssignVariableOp", "EmptyTensorList", "MergeV2Checkpoints", "ReadVariableOp", "ResourceGather", "RestoreV2", "SaveV2", "ShardedFilename", "StatefulPartitionedCall", "StaticRegexFullMatch", "VarHandleOp", ] def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = SavedModel() lowercase = [] with open(os.path.join(lowerCAmelCase__ , '''utils''' , '''tf_ops''' , '''onnx.json''' ) ) as f: lowercase = json.load(lowerCAmelCase__ )['''opsets'''] for i in range(1 , opset + 1 ): onnx_ops.extend(onnx_opsets[str(lowerCAmelCase__ )] ) with open(lowerCAmelCase__ , '''rb''' ) as f: saved_model.ParseFromString(f.read() ) lowercase = set() # Iterate over every metagraph in case there is more than one (a saved model can contain multiple graphs) for meta_graph in saved_model.meta_graphs: # Add operations in the graph definition model_op_names.update(node.op for node in meta_graph.graph_def.node ) # Go through the functions in the graph definition for func in meta_graph.graph_def.library.function: # Add operations in each function model_op_names.update(node.op for node in func.node_def ) # Convert to list, sorted if you want lowercase = sorted(lowerCAmelCase__ ) lowercase = [] for op in model_op_names: if op not in onnx_ops and op not in INTERNAL_OPS: incompatible_ops.append(lowerCAmelCase__ ) if strict and len(lowerCAmelCase__ ) > 0: raise Exception(f'Found the following incompatible ops for the opset {opset}:\n' + incompatible_ops ) elif len(lowerCAmelCase__ ) > 0: print(f'Found the following incompatible ops for the opset {opset}:' ) print(*lowerCAmelCase__ , sep='''\n''' ) else: print(f'The saved model {saved_model_path} can properly be converted with ONNX.' ) if __name__ == "__main__": lowercase__ :Optional[Any] = argparse.ArgumentParser() parser.add_argument("--saved_model_path", help="Path of the saved model to check (the .pb file).") parser.add_argument( "--opset", default=12, type=int, help="The ONNX opset against which the model has to be tested." ) parser.add_argument( "--framework", choices=["onnx"], default="onnx", help="Frameworks against which to test the saved model." ) parser.add_argument( "--strict", action="store_true", help="Whether make the checking strict (raise errors) or not (raise warnings)" ) lowercase__ :int = parser.parse_args() if args.framework == "onnx": onnx_compliancy(args.saved_model_path, args.strict, args.opset)
101
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case__ : Optional[Any] = {'configuration_ibert': ['IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'IBertConfig', 'IBertOnnxConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ : List[str] = [ 'IBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'IBertForMaskedLM', 'IBertForMultipleChoice', 'IBertForQuestionAnswering', 'IBertForSequenceClassification', 'IBertForTokenClassification', 'IBertModel', 'IBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys snake_case__ : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
117
0
"""simple docstring""" from typing import List, Optional, Tuple, Union import torch from ...utils import logging, randn_tensor from ..pipeline_utils import AudioPipelineOutput, DiffusionPipeline A_ = logging.get_logger(__name__) # pylint: disable=invalid-name class __SCREAMING_SNAKE_CASE ( a__ ): def __init__( self : int , snake_case : List[Any] , snake_case : str ): '''simple docstring''' super().__init__() self.register_modules(unet=snake_case , scheduler=snake_case ) @torch.no_grad() def __call__( self : List[str] , snake_case : int = 1 , snake_case : int = 100 , snake_case : Optional[Union[torch.Generator, List[torch.Generator]]] = None , snake_case : Optional[float] = None , snake_case : bool = True , ): '''simple docstring''' if audio_length_in_s is None: A__ : Optional[Any] = self.unet.config.sample_size / self.unet.config.sample_rate A__ : List[Any] = audio_length_in_s * self.unet.config.sample_rate A__ : Tuple = 2 ** len(self.unet.up_blocks ) if sample_size < 3 * down_scale_factor: raise ValueError( F'{audio_length_in_s} is too small. Make sure it\'s bigger or equal to' F' {3 * down_scale_factor / self.unet.config.sample_rate}.' ) A__ : List[str] = int(snake_case ) if sample_size % down_scale_factor != 0: A__ : Any = ( (audio_length_in_s * self.unet.config.sample_rate) // down_scale_factor + 1 ) * down_scale_factor logger.info( F'{audio_length_in_s} is increased to {sample_size / self.unet.config.sample_rate} so that it can be handled' F' by the model. It will be cut to {original_sample_size / self.unet.config.sample_rate} after the denoising' """ process.""" ) A__ : Optional[int] = int(snake_case ) A__ : str = next(iter(self.unet.parameters() ) ).dtype A__ : Tuple = (batch_size, self.unet.config.in_channels, sample_size) if isinstance(snake_case , snake_case ) and len(snake_case ) != batch_size: raise ValueError( F'You have passed a list of generators of length {len(snake_case )}, but requested an effective batch' F' size of {batch_size}. Make sure the batch size matches the length of the generators.' ) A__ : List[str] = randn_tensor(snake_case , generator=snake_case , device=self.device , dtype=snake_case ) # set step values self.scheduler.set_timesteps(snake_case , device=audio.device ) A__ : Tuple = self.scheduler.timesteps.to(snake_case ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output A__ : Tuple = self.unet(snake_case , snake_case ).sample # 2. compute previous image: x_t -> t_t-1 A__ : Union[str, Any] = self.scheduler.step(snake_case , snake_case , snake_case ).prev_sample A__ : Tuple = audio.clamp(-1 , 1 ).float().cpu().numpy() A__ : Union[str, Any] = audio[:, :, :original_sample_size] if not return_dict: return (audio,) return AudioPipelineOutput(audios=snake_case )
370
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) def _lowerCAmelCase ( UpperCAmelCase__ : Union[str, Any], UpperCAmelCase__ : Tuple=False ) ->str: A__ : Optional[int] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'blocks.{i}.norm1.weight', f'deit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((f'blocks.{i}.norm1.bias', f'deit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append((f'blocks.{i}.attn.proj.weight', f'deit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append((f'blocks.{i}.attn.proj.bias', f'deit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((f'blocks.{i}.norm2.weight', f'deit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((f'blocks.{i}.norm2.bias', f'deit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((f'blocks.{i}.mlp.fc1.weight', f'deit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((f'blocks.{i}.mlp.fc1.bias', f'deit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((f'blocks.{i}.mlp.fc2.weight', f'deit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((f'blocks.{i}.mlp.fc2.bias', f'deit.encoder.layer.{i}.output.dense.bias') ) # projection layer + position embeddings rename_keys.extend( [ ("""cls_token""", """deit.embeddings.cls_token"""), ("""dist_token""", """deit.embeddings.distillation_token"""), ("""patch_embed.proj.weight""", """deit.embeddings.patch_embeddings.projection.weight"""), ("""patch_embed.proj.bias""", """deit.embeddings.patch_embeddings.projection.bias"""), ("""pos_embed""", """deit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""norm.weight""", """layernorm.weight"""), ("""norm.bias""", """layernorm.bias"""), ("""pre_logits.fc.weight""", """pooler.dense.weight"""), ("""pre_logits.fc.bias""", """pooler.dense.bias"""), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" A__ : Optional[int] = [(pair[0], pair[1][4:]) if pair[1].startswith("""deit""" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("""norm.weight""", """deit.layernorm.weight"""), ("""norm.bias""", """deit.layernorm.bias"""), ("""head.weight""", """cls_classifier.weight"""), ("""head.bias""", """cls_classifier.bias"""), ("""head_dist.weight""", """distillation_classifier.weight"""), ("""head_dist.bias""", """distillation_classifier.bias"""), ] ) return rename_keys def _lowerCAmelCase ( UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : Tuple, UpperCAmelCase__ : List[Any]=False ) ->str: for i in range(config.num_hidden_layers ): if base_model: A__ : Any = """""" else: A__ : Tuple = """deit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A__ : Any = state_dict.pop(f'blocks.{i}.attn.qkv.weight' ) A__ : Tuple = state_dict.pop(f'blocks.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict A__ : List[Any] = in_proj_weight[ : config.hidden_size, : ] A__ : str = in_proj_bias[: config.hidden_size] A__ : Any = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A__ : Dict = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A__ : Optional[Any] = in_proj_weight[ -config.hidden_size :, : ] A__ : Any = in_proj_bias[-config.hidden_size :] def _lowerCAmelCase ( UpperCAmelCase__ : List[Any], UpperCAmelCase__ : List[Any], UpperCAmelCase__ : Union[str, Any] ) ->Any: A__ : int = dct.pop(UpperCAmelCase__ ) A__ : Tuple = val def _lowerCAmelCase ( ) ->List[Any]: A__ : Optional[int] = """http://images.cocodataset.org/val2017/000000039769.jpg""" A__ : int = Image.open(requests.get(UpperCAmelCase__, stream=UpperCAmelCase__ ).raw ) return im @torch.no_grad() def _lowerCAmelCase ( UpperCAmelCase__ : Dict, UpperCAmelCase__ : Any ) ->Tuple: A__ : List[Any] = DeiTConfig() # all deit models have fine-tuned heads A__ : Tuple = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size A__ : str = 1_0_0_0 A__ : List[str] = """huggingface/label-files""" A__ : Dict = """imagenet-1k-id2label.json""" A__ : List[str] = json.load(open(hf_hub_download(UpperCAmelCase__, UpperCAmelCase__, repo_type="""dataset""" ), """r""" ) ) A__ : Dict = {int(UpperCAmelCase__ ): v for k, v in idalabel.items()} A__ : Optional[int] = idalabel A__ : Dict = {v: k for k, v in idalabel.items()} A__ : List[str] = int(deit_name[-6:-4] ) A__ : str = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("""tiny""" ): A__ : List[str] = 1_9_2 A__ : int = 7_6_8 A__ : List[Any] = 1_2 A__ : Dict = 3 elif deit_name[9:].startswith("""small""" ): A__ : List[Any] = 3_8_4 A__ : List[str] = 1_5_3_6 A__ : Any = 1_2 A__ : Union[str, Any] = 6 if deit_name[9:].startswith("""base""" ): pass elif deit_name[4:].startswith("""large""" ): A__ : int = 1_0_2_4 A__ : str = 4_0_9_6 A__ : Any = 2_4 A__ : int = 1_6 # load original model from timm A__ : Dict = timm.create_model(UpperCAmelCase__, pretrained=UpperCAmelCase__ ) timm_model.eval() # load state_dict of original model, remove and rename some keys A__ : Tuple = timm_model.state_dict() A__ : str = create_rename_keys(UpperCAmelCase__, UpperCAmelCase__ ) for src, dest in rename_keys: rename_key(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) read_in_q_k_v(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) # load HuggingFace model A__ : str = DeiTForImageClassificationWithTeacher(UpperCAmelCase__ ).eval() model.load_state_dict(UpperCAmelCase__ ) # Check outputs on an image, prepared by DeiTImageProcessor A__ : int = int( (2_5_6 / 2_2_4) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 A__ : Any = DeiTImageProcessor(size=UpperCAmelCase__, crop_size=config.image_size ) A__ : Union[str, Any] = image_processor(images=prepare_img(), return_tensors="""pt""" ) A__ : Optional[Any] = encoding["""pixel_values"""] A__ : Union[str, Any] = model(UpperCAmelCase__ ) A__ : Union[str, Any] = timm_model(UpperCAmelCase__ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(UpperCAmelCase__, outputs.logits, atol=1e-3 ) Path(UpperCAmelCase__ ).mkdir(exist_ok=UpperCAmelCase__ ) print(f'Saving model {deit_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(UpperCAmelCase__ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(UpperCAmelCase__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--deit_name''', default='''vit_deit_base_distilled_patch16_224''', type=str, help='''Name of the DeiT timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) A_ = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
296
0
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class lowercase ( unittest.TestCase ): """simple docstring""" def __init__( self , UpperCamelCase_ , UpperCamelCase_=7 , UpperCamelCase_=3 , UpperCamelCase_=18 , UpperCamelCase_=30 , UpperCamelCase_=400 , UpperCamelCase_=True , UpperCamelCase_=None , UpperCamelCase_=True , ): '''simple docstring''' UpperCamelCase__ :Any = size if size is not None else {'''height''': 18, '''width''': 18} UpperCamelCase__ :Union[str, Any] = parent UpperCamelCase__ :Tuple = batch_size UpperCamelCase__ :int = num_channels UpperCamelCase__ :Optional[Any] = image_size UpperCamelCase__ :List[str] = min_resolution UpperCamelCase__ :Optional[int] = max_resolution UpperCamelCase__ :Optional[Any] = do_resize UpperCamelCase__ :Optional[int] = size UpperCamelCase__ :List[Any] = apply_ocr def lowerCAmelCase__ ( self ): '''simple docstring''' return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class lowercase ( A__ , unittest.TestCase ): """simple docstring""" _a = LayoutLMvaImageProcessor if is_pytesseract_available() else None def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :List[Any] = LayoutLMvaImageProcessingTester(self ) @property def lowerCAmelCase__ ( self ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Tuple = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCamelCase_ , '''do_resize''' ) ) self.assertTrue(hasattr(UpperCamelCase_ , '''size''' ) ) self.assertTrue(hasattr(UpperCamelCase_ , '''apply_ocr''' ) ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Any = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''height''': 18, '''width''': 18} ) UpperCamelCase__ :Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {'''height''': 42, '''width''': 42} ) def lowerCAmelCase__ ( self ): '''simple docstring''' pass def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCamelCase__ :Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , Image.Image ) # Test not batched input UpperCamelCase__ :List[str] = image_processing(image_inputs[0] , return_tensors='''pt''' ) self.assertEqual( encoding.pixel_values.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) self.assertIsInstance(encoding.words , UpperCamelCase_ ) self.assertIsInstance(encoding.boxes , UpperCamelCase_ ) # Test batched UpperCamelCase__ :int = image_processing(UpperCamelCase_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCamelCase__ :Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_ , numpify=UpperCamelCase_ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , np.ndarray ) # Test not batched input UpperCamelCase__ :Optional[int] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched UpperCamelCase__ :Tuple = image_processing(UpperCamelCase_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Any = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCamelCase__ :List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_ , torchify=UpperCamelCase_ ) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , torch.Tensor ) # Test not batched input UpperCamelCase__ :Union[str, Any] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched UpperCamelCase__ :Dict = image_processing(UpperCamelCase_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Union[str, Any] = LayoutLMvaImageProcessor() from datasets import load_dataset UpperCamelCase__ :int = load_dataset('''hf-internal-testing/fixtures_docvqa''' , split='''test''' ) UpperCamelCase__ :int = Image.open(ds[0]['''file'''] ).convert('''RGB''' ) UpperCamelCase__ :str = image_processing(UpperCamelCase_ , return_tensors='''pt''' ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224) ) self.assertEqual(len(encoding.words ) , len(encoding.boxes ) ) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 UpperCamelCase__ :Tuple = [['''11:14''', '''to''', '''11:39''', '''a.m''', '''11:39''', '''to''', '''11:44''', '''a.m.''', '''11:44''', '''a.m.''', '''to''', '''12:25''', '''p.m.''', '''12:25''', '''to''', '''12:58''', '''p.m.''', '''12:58''', '''to''', '''4:00''', '''p.m.''', '''2:00''', '''to''', '''5:00''', '''p.m.''', '''Coffee''', '''Break''', '''Coffee''', '''will''', '''be''', '''served''', '''for''', '''men''', '''and''', '''women''', '''in''', '''the''', '''lobby''', '''adjacent''', '''to''', '''exhibit''', '''area.''', '''Please''', '''move''', '''into''', '''exhibit''', '''area.''', '''(Exhibits''', '''Open)''', '''TRRF''', '''GENERAL''', '''SESSION''', '''(PART''', '''|)''', '''Presiding:''', '''Lee''', '''A.''', '''Waller''', '''TRRF''', '''Vice''', '''President''', '''“Introductory''', '''Remarks”''', '''Lee''', '''A.''', '''Waller,''', '''TRRF''', '''Vice''', '''Presi-''', '''dent''', '''Individual''', '''Interviews''', '''with''', '''TRRF''', '''Public''', '''Board''', '''Members''', '''and''', '''Sci-''', '''entific''', '''Advisory''', '''Council''', '''Mem-''', '''bers''', '''Conducted''', '''by''', '''TRRF''', '''Treasurer''', '''Philip''', '''G.''', '''Kuehn''', '''to''', '''get''', '''answers''', '''which''', '''the''', '''public''', '''refrigerated''', '''warehousing''', '''industry''', '''is''', '''looking''', '''for.''', '''Plus''', '''questions''', '''from''', '''the''', '''floor.''', '''Dr.''', '''Emil''', '''M.''', '''Mrak,''', '''University''', '''of''', '''Cal-''', '''ifornia,''', '''Chairman,''', '''TRRF''', '''Board;''', '''Sam''', '''R.''', '''Cecil,''', '''University''', '''of''', '''Georgia''', '''College''', '''of''', '''Agriculture;''', '''Dr.''', '''Stanley''', '''Charm,''', '''Tufts''', '''University''', '''School''', '''of''', '''Medicine;''', '''Dr.''', '''Robert''', '''H.''', '''Cotton,''', '''ITT''', '''Continental''', '''Baking''', '''Company;''', '''Dr.''', '''Owen''', '''Fennema,''', '''University''', '''of''', '''Wis-''', '''consin;''', '''Dr.''', '''Robert''', '''E.''', '''Hardenburg,''', '''USDA.''', '''Questions''', '''and''', '''Answers''', '''Exhibits''', '''Open''', '''Capt.''', '''Jack''', '''Stoney''', '''Room''', '''TRRF''', '''Scientific''', '''Advisory''', '''Council''', '''Meeting''', '''Ballroom''', '''Foyer''']] # noqa: E231 UpperCamelCase__ :Dict = [[[141, 57, 214, 69], [228, 58, 252, 69], [141, 75, 216, 88], [230, 79, 280, 88], [142, 260, 218, 273], [230, 261, 255, 273], [143, 279, 218, 290], [231, 282, 290, 291], [143, 342, 218, 354], [231, 345, 289, 355], [202, 362, 227, 373], [143, 379, 220, 392], [231, 382, 291, 394], [144, 714, 220, 726], [231, 715, 256, 726], [144, 732, 220, 745], [232, 736, 291, 747], [144, 769, 218, 782], [231, 770, 256, 782], [141, 788, 202, 801], [215, 791, 274, 804], [143, 826, 204, 838], [215, 826, 240, 838], [142, 844, 202, 857], [215, 847, 274, 859], [334, 57, 427, 69], [440, 57, 522, 69], [369, 75, 461, 88], [469, 75, 516, 88], [528, 76, 562, 88], [570, 76, 667, 88], [675, 75, 711, 87], [721, 79, 778, 88], [789, 75, 840, 88], [369, 97, 470, 107], [484, 94, 507, 106], [518, 94, 562, 107], [576, 94, 655, 110], [668, 94, 792, 109], [804, 95, 829, 107], [369, 113, 465, 125], [477, 116, 547, 125], [562, 113, 658, 125], [671, 116, 748, 125], [761, 113, 811, 125], [369, 131, 465, 143], [477, 133, 548, 143], [563, 130, 698, 145], [710, 130, 802, 146], [336, 171, 412, 183], [423, 171, 572, 183], [582, 170, 716, 184], [728, 171, 817, 187], [829, 171, 844, 186], [338, 197, 482, 212], [507, 196, 557, 209], [569, 196, 595, 208], [610, 196, 702, 209], [505, 214, 583, 226], [595, 214, 656, 227], [670, 215, 807, 227], [335, 259, 543, 274], [556, 259, 708, 272], [372, 279, 422, 291], [435, 279, 460, 291], [474, 279, 574, 292], [587, 278, 664, 291], [676, 278, 738, 291], [751, 279, 834, 291], [372, 298, 434, 310], [335, 341, 483, 354], [497, 341, 655, 354], [667, 341, 728, 354], [740, 341, 825, 354], [335, 360, 430, 372], [442, 360, 534, 372], [545, 359, 687, 372], [697, 360, 754, 372], [765, 360, 823, 373], [334, 378, 428, 391], [440, 378, 577, 394], [590, 378, 705, 391], [720, 378, 801, 391], [334, 397, 400, 409], [370, 416, 529, 429], [544, 416, 576, 432], [587, 416, 665, 428], [677, 416, 814, 429], [372, 435, 452, 450], [465, 434, 495, 447], [511, 434, 600, 447], [611, 436, 637, 447], [649, 436, 694, 451], [705, 438, 824, 447], [369, 453, 452, 466], [464, 454, 509, 466], [522, 453, 611, 469], [625, 453, 792, 469], [370, 472, 556, 488], [570, 472, 684, 487], [697, 472, 718, 485], [732, 472, 835, 488], [369, 490, 411, 503], [425, 490, 484, 503], [496, 490, 635, 506], [645, 490, 707, 503], [718, 491, 761, 503], [771, 490, 840, 503], [336, 510, 374, 521], [388, 510, 447, 522], [460, 510, 489, 521], [503, 510, 580, 522], [592, 509, 736, 525], [745, 509, 770, 522], [781, 509, 840, 522], [338, 528, 434, 541], [448, 528, 596, 541], [609, 527, 687, 540], [700, 528, 792, 541], [336, 546, 397, 559], [407, 546, 431, 559], [443, 546, 525, 560], [537, 546, 680, 562], [688, 546, 714, 559], [722, 546, 837, 562], [336, 565, 449, 581], [461, 565, 485, 577], [497, 565, 665, 581], [681, 565, 718, 577], [732, 565, 837, 580], [337, 584, 438, 597], [452, 583, 521, 596], [535, 584, 677, 599], [690, 583, 787, 596], [801, 583, 825, 596], [338, 602, 478, 615], [492, 602, 530, 614], [543, 602, 638, 615], [650, 602, 676, 614], [688, 602, 788, 615], [802, 602, 843, 614], [337, 621, 502, 633], [516, 621, 615, 637], [629, 621, 774, 636], [789, 621, 827, 633], [337, 639, 418, 652], [432, 640, 571, 653], [587, 639, 731, 655], [743, 639, 769, 652], [780, 639, 841, 652], [338, 658, 440, 673], [455, 658, 491, 670], [508, 658, 602, 671], [616, 658, 638, 670], [654, 658, 835, 674], [337, 677, 429, 689], [337, 714, 482, 726], [495, 714, 548, 726], [561, 714, 683, 726], [338, 770, 461, 782], [474, 769, 554, 785], [489, 788, 562, 803], [576, 788, 643, 801], [656, 787, 751, 804], [764, 788, 844, 801], [334, 825, 421, 838], [430, 824, 574, 838], [584, 824, 723, 841], [335, 844, 450, 857], [464, 843, 583, 860], [628, 862, 755, 875], [769, 861, 848, 878]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words , UpperCamelCase_ ) self.assertListEqual(encoding.boxes , UpperCamelCase_ ) # with apply_OCR = False UpperCamelCase__ :List[str] = LayoutLMvaImageProcessor(apply_ocr=UpperCamelCase_ ) UpperCamelCase__ :Tuple = image_processing(UpperCamelCase_ , return_tensors='''pt''' ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224) )
97
'''simple docstring''' import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot import BlenderbotTokenizer if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __snake_case =logging.get_logger(__name__) __snake_case ={ """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } __snake_case ={ """vocab_file""": {"""facebook/blenderbot-3B""": """https://huggingface.co/facebook/blenderbot-3B/resolve/main/vocab.json"""}, """merges_file""": {"""facebook/blenderbot-3B""": """https://huggingface.co/facebook/blenderbot-3B/resolve/main/merges.txt"""}, """tokenizer_config_file""": { """facebook/blenderbot-3B""": """https://huggingface.co/facebook/blenderbot-3B/resolve/main/tokenizer_config.json""" }, } __snake_case ={"""facebook/blenderbot-3B""": 128} class UpperCAmelCase_ ( __lowercase ): lowerCamelCase : List[Any] = VOCAB_FILES_NAMES lowerCamelCase : Optional[int] = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Optional[Any] = ['''input_ids''', '''attention_mask'''] lowerCamelCase : List[Any] = BlenderbotTokenizer def __init__( self : Union[str, Any] , UpperCAmelCase__ : Tuple=None , UpperCAmelCase__ : Tuple=None , UpperCAmelCase__ : List[Any]=None , UpperCAmelCase__ : str="replace" , UpperCAmelCase__ : Dict="<s>" , UpperCAmelCase__ : Tuple="</s>" , UpperCAmelCase__ : Optional[Any]="</s>" , UpperCAmelCase__ : Any="<s>" , UpperCAmelCase__ : List[str]="<unk>" , UpperCAmelCase__ : int="<pad>" , UpperCAmelCase__ : Union[str, Any]="<mask>" , UpperCAmelCase__ : str=False , UpperCAmelCase__ : Union[str, Any]=True , **UpperCAmelCase__ : Optional[int] , ) -> int: super().__init__( UpperCAmelCase__ , UpperCAmelCase__ , tokenizer_file=UpperCAmelCase__ , errors=UpperCAmelCase__ , bos_token=UpperCAmelCase__ , eos_token=UpperCAmelCase__ , sep_token=UpperCAmelCase__ , cls_token=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , mask_token=UpperCAmelCase__ , add_prefix_space=UpperCAmelCase__ , trim_offsets=UpperCAmelCase__ , **UpperCAmelCase__ , ) lowerCAmelCase = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('add_prefix_space' , UpperCAmelCase__ ) != add_prefix_space: lowerCAmelCase = getattr(UpperCAmelCase__ , pre_tok_state.pop('type' ) ) lowerCAmelCase = add_prefix_space lowerCAmelCase = pre_tok_class(**UpperCAmelCase__ ) lowerCAmelCase = add_prefix_space lowerCAmelCase = 'post_processor' lowerCAmelCase = getattr(self.backend_tokenizer , UpperCAmelCase__ , UpperCAmelCase__ ) if tokenizer_component_instance: lowerCAmelCase = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowerCAmelCase = tuple(state['sep'] ) if "cls" in state: lowerCAmelCase = tuple(state['cls'] ) lowerCAmelCase = False if state.get('add_prefix_space' , UpperCAmelCase__ ) != add_prefix_space: lowerCAmelCase = add_prefix_space lowerCAmelCase = True if state.get('trim_offsets' , UpperCAmelCase__ ) != trim_offsets: lowerCAmelCase = trim_offsets lowerCAmelCase = True if changes_to_apply: lowerCAmelCase = getattr(UpperCAmelCase__ , state.pop('type' ) ) lowerCAmelCase = component_class(**UpperCAmelCase__ ) setattr(self.backend_tokenizer , UpperCAmelCase__ , UpperCAmelCase__ ) @property # Copied from transformers.models.roberta.tokenization_roberta_fast.RobertaTokenizerFast.mask_token with Roberta->Blenderbot, RoBERTa->Blenderbot def __UpperCAmelCase ( self : Union[str, Any] ) -> str: if self._mask_token is None: if self.verbose: logger.error('Using mask_token, but it is not set yet.' ) return None return str(self._mask_token ) @mask_token.setter def __UpperCAmelCase ( self : int , UpperCAmelCase__ : Optional[Any] ) -> Tuple: lowerCAmelCase = AddedToken(UpperCAmelCase__ , lstrip=UpperCAmelCase__ , rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else value lowerCAmelCase = value def __UpperCAmelCase ( self : Optional[Any] , *UpperCAmelCase__ : Union[str, Any] , **UpperCAmelCase__ : List[str] ) -> BatchEncoding: lowerCAmelCase = kwargs.get('is_split_into_words' , UpperCAmelCase__ ) assert self.add_prefix_space or not is_split_into_words, ( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*UpperCAmelCase__ , **UpperCAmelCase__ ) def __UpperCAmelCase ( self : List[str] , *UpperCAmelCase__ : str , **UpperCAmelCase__ : List[str] ) -> BatchEncoding: lowerCAmelCase = kwargs.get('is_split_into_words' , UpperCAmelCase__ ) assert self.add_prefix_space or not is_split_into_words, ( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._encode_plus(*UpperCAmelCase__ , **UpperCAmelCase__ ) def __UpperCAmelCase ( self : str , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ) -> Tuple[str]: lowerCAmelCase = self._tokenizer.model.save(UpperCAmelCase__ , name=UpperCAmelCase__ ) return tuple(UpperCAmelCase__ ) def __UpperCAmelCase ( self : Optional[int] , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ) -> List[int]: lowerCAmelCase = [self.sep_token_id] lowerCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __UpperCAmelCase ( self : List[str] , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ) -> Any: return token_ids_a + [self.eos_token_id] def __UpperCAmelCase ( self : Union[str, Any] , UpperCAmelCase__ : "Conversation" ) -> List[int]: lowerCAmelCase = [] for is_user, text in conversation.iter_texts(): if is_user: # We need to space prefix as it's being done within blenderbot inputs.append(' ' + text ) else: # Generated responses should contain them already. inputs.append(UpperCAmelCase__ ) lowerCAmelCase = ' '.join(UpperCAmelCase__ ) lowerCAmelCase = self.encode(UpperCAmelCase__ ) if len(UpperCAmelCase__ ) > self.model_max_length: lowerCAmelCase = input_ids[-self.model_max_length :] logger.warning(F'''Trimmed input from conversation as it was longer than {self.model_max_length} tokens.''' ) return input_ids
4
0
'''simple docstring''' import os from pathlib import Path import numpy as np import pytest from pack_dataset import pack_data_dir from parameterized import parameterized from save_len_file import save_len_file from torch.utils.data import DataLoader from transformers import AutoTokenizer from transformers.models.mbart.modeling_mbart import shift_tokens_right from transformers.testing_utils import TestCasePlus, slow from utils import FAIRSEQ_AVAILABLE, DistributedSortishSampler, LegacySeqaSeqDataset, SeqaSeqDataset UpperCamelCase__ = '''bert-base-cased''' UpperCamelCase__ = '''google/pegasus-xsum''' UpperCamelCase__ = [''' Sam ate lunch today.''', '''Sams lunch ingredients.'''] UpperCamelCase__ = ['''A very interesting story about what I ate for lunch.''', '''Avocado, celery, turkey, coffee'''] UpperCamelCase__ = '''patrickvonplaten/t5-tiny-random''' UpperCamelCase__ = '''sshleifer/bart-tiny-random''' UpperCamelCase__ = '''sshleifer/tiny-mbart''' UpperCamelCase__ = '''sshleifer/tiny-marian-en-de''' def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ) -> str: UpperCAmelCase__ : Optional[Any] = '''\n'''.join(__a ) Path(__a ).open('''w''' ).writelines(__a ) def a__ ( lowerCAmelCase__ ) -> List[str]: for split in ["train", "val", "test"]: _dump_articles(os.path.join(__a , F"""{split}.source""" ) , __a ) _dump_articles(os.path.join(__a , F"""{split}.target""" ) , __a ) return tmp_dir class lowerCamelCase_ ( __lowercase ): @parameterized.expand( [ MBART_TINY, MARIAN_TINY, T5_TINY, BART_TINY, PEGASUS_XSUM, ] , ) @slow def lowercase_ ( self : Tuple , _A : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : Any = AutoTokenizer.from_pretrained(_a ) UpperCAmelCase__ : Tuple = make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) UpperCAmelCase__ : Tuple = max(len(tokenizer.encode(_a ) ) for a in ARTICLES ) UpperCAmelCase__ : List[str] = max(len(tokenizer.encode(_a ) ) for a in SUMMARIES ) UpperCAmelCase__ : str = 4 UpperCAmelCase__ : str = 8 assert max_len_target > max_src_len # Will be truncated assert max_len_source > max_src_len # Will be truncated UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = '''ro_RO''', '''de_DE''' # ignored for all but mbart, but never causes error. UpperCAmelCase__ : Dict = SeqaSeqDataset( _a , data_dir=_a , type_path='''train''' , max_source_length=_a , max_target_length=_a , src_lang=_a , tgt_lang=_a , ) UpperCAmelCase__ : Union[str, Any] = DataLoader(_a , batch_size=2 , collate_fn=train_dataset.collate_fn ) for batch in dataloader: assert isinstance(_a , _a ) assert batch["attention_mask"].shape == batch["input_ids"].shape # show that articles were trimmed. assert batch["input_ids"].shape[1] == max_src_len # show that targets are the same len assert batch["labels"].shape[1] == max_tgt_len if tok_name != MBART_TINY: continue # check language codes in correct place UpperCAmelCase__ : str = shift_tokens_right(batch['''labels'''] , tokenizer.pad_token_id ) assert batch["decoder_input_ids"][0, 0].item() == tokenizer.lang_code_to_id[tgt_lang] assert batch["decoder_input_ids"][0, -1].item() == tokenizer.eos_token_id assert batch["input_ids"][0, -2].item() == tokenizer.eos_token_id assert batch["input_ids"][0, -1].item() == tokenizer.lang_code_to_id[src_lang] break # No need to test every batch @parameterized.expand([BART_TINY, BERT_BASE_CASED] ) def lowercase_ ( self : Any , _A : List[str] ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = AutoTokenizer.from_pretrained(_a ) UpperCAmelCase__ : Optional[Any] = make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) UpperCAmelCase__ : Dict = max(len(tokenizer.encode(_a ) ) for a in ARTICLES ) UpperCAmelCase__ : Optional[int] = max(len(tokenizer.encode(_a ) ) for a in SUMMARIES ) UpperCAmelCase__ : List[Any] = 4 UpperCAmelCase__ : Optional[Any] = LegacySeqaSeqDataset( _a , data_dir=_a , type_path='''train''' , max_source_length=20 , max_target_length=_a , ) UpperCAmelCase__ : Optional[int] = DataLoader(_a , batch_size=2 , collate_fn=train_dataset.collate_fn ) for batch in dataloader: assert batch["attention_mask"].shape == batch["input_ids"].shape # show that articles were trimmed. assert batch["input_ids"].shape[1] == max_len_source assert 20 >= batch["input_ids"].shape[1] # trimmed significantly # show that targets were truncated assert batch["labels"].shape[1] == trunc_target # Truncated assert max_len_target > trunc_target # Truncated break # No need to test every batch def lowercase_ ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = AutoTokenizer.from_pretrained('''facebook/mbart-large-cc25''' ) UpperCAmelCase__ : Optional[Any] = Path(make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) ) UpperCAmelCase__ : List[Any] = tmp_dir.joinpath('''train.source''' ).open().readlines() UpperCAmelCase__ : str = Path(make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) ) pack_data_dir(_a , _a , 128 , _a ) UpperCAmelCase__ : List[Any] = {x.name for x in tmp_dir.iterdir()} UpperCAmelCase__ : int = {x.name for x in save_dir.iterdir()} UpperCAmelCase__ : Any = save_dir.joinpath('''train.source''' ).open().readlines() # orig: [' Sam ate lunch today.\n', 'Sams lunch ingredients.'] # desired_packed: [' Sam ate lunch today.\n Sams lunch ingredients.'] assert len(_a ) < len(_a ) assert len(_a ) == 1 assert len(packed_examples[0] ) == sum(len(_a ) for x in orig_examples ) assert orig_paths == new_paths @pytest.mark.skipif(not FAIRSEQ_AVAILABLE , reason='''This test requires fairseq''' ) def lowercase_ ( self : str ): '''simple docstring''' if not FAIRSEQ_AVAILABLE: return UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self._get_dataset(max_len=64 ) UpperCAmelCase__ : Union[str, Any] = 64 UpperCAmelCase__ : Optional[int] = ds.make_dynamic_sampler(_a , required_batch_size_multiple=_a ) UpperCAmelCase__ : Union[str, Any] = [len(_a ) for x in batch_sampler] assert len(set(_a ) ) > 1 # it's not dynamic batch size if every batch is the same length assert sum(_a ) == len(_a ) # no dropped or added examples UpperCAmelCase__ : str = DataLoader(_a , batch_sampler=_a , collate_fn=ds.collate_fn , num_workers=2 ) UpperCAmelCase__ : Dict = [] UpperCAmelCase__ : Tuple = [] for batch in data_loader: UpperCAmelCase__ : Dict = batch['''input_ids'''].shape UpperCAmelCase__ : str = src_shape[0] assert bs % required_batch_size_multiple == 0 or bs < required_batch_size_multiple UpperCAmelCase__ : str = np.product(batch['''input_ids'''].shape ) num_src_per_batch.append(_a ) if num_src_tokens > (max_tokens * 1.1): failures.append(_a ) assert num_src_per_batch[0] == max(_a ) if failures: raise AssertionError(f"""too many tokens in {len(_a )} batches""" ) def lowercase_ ( self : Dict ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Union[str, Any] = self._get_dataset(max_len=512 ) UpperCAmelCase__ : List[str] = 2 UpperCAmelCase__ : Any = ds.make_sortish_sampler(_a , shuffle=_a ) UpperCAmelCase__ : Any = DataLoader(_a , batch_size=_a , collate_fn=ds.collate_fn , num_workers=2 ) UpperCAmelCase__ : List[Any] = DataLoader(_a , batch_size=_a , collate_fn=ds.collate_fn , num_workers=2 , sampler=_a ) UpperCAmelCase__ : str = tokenizer.pad_token_id def count_pad_tokens(_A : Any , _A : Any="input_ids" ): return [batch[k].eq(_a ).sum().item() for batch in data_loader] assert sum(count_pad_tokens(_a , k='''labels''' ) ) < sum(count_pad_tokens(_a , k='''labels''' ) ) assert sum(count_pad_tokens(_a ) ) < sum(count_pad_tokens(_a ) ) assert len(_a ) == len(_a ) def lowercase_ ( self : List[Any] , _A : List[str]=1_000 , _A : Dict=128 ): '''simple docstring''' if os.getenv('''USE_REAL_DATA''' , _a ): UpperCAmelCase__ : List[Any] = '''examples/seq2seq/wmt_en_ro''' UpperCAmelCase__ : Tuple = max_len * 2 * 64 if not Path(_a ).joinpath('''train.len''' ).exists(): save_len_file(_a , _a ) else: UpperCAmelCase__ : Tuple = '''examples/seq2seq/test_data/wmt_en_ro''' UpperCAmelCase__ : Union[str, Any] = max_len * 4 save_len_file(_a , _a ) UpperCAmelCase__ : int = AutoTokenizer.from_pretrained(_a ) UpperCAmelCase__ : str = SeqaSeqDataset( _a , data_dir=_a , type_path='''train''' , max_source_length=_a , max_target_length=_a , n_obs=_a , ) return ds, max_tokens, tokenizer def lowercase_ ( self : Any ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : int = self._get_dataset() UpperCAmelCase__ : Optional[int] = set(DistributedSortishSampler(_a , 256 , num_replicas=2 , rank=0 , add_extra_examples=_a ) ) UpperCAmelCase__ : int = set(DistributedSortishSampler(_a , 256 , num_replicas=2 , rank=1 , add_extra_examples=_a ) ) assert idsa.intersection(_a ) == set() @parameterized.expand( [ MBART_TINY, MARIAN_TINY, T5_TINY, BART_TINY, PEGASUS_XSUM, ] , ) def lowercase_ ( self : Dict , _A : str ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = AutoTokenizer.from_pretrained(_a , use_fast=_a ) if tok_name == MBART_TINY: UpperCAmelCase__ : Optional[Any] = SeqaSeqDataset( _a , data_dir=make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) , type_path='''train''' , max_source_length=4 , max_target_length=8 , src_lang='''EN''' , tgt_lang='''FR''' , ) UpperCAmelCase__ : Tuple = train_dataset.dataset_kwargs assert "src_lang" in kwargs and "tgt_lang" in kwargs else: UpperCAmelCase__ : List[str] = SeqaSeqDataset( _a , data_dir=make_test_data_dir(tmp_dir=self.get_auto_remove_tmp_dir() ) , type_path='''train''' , max_source_length=4 , max_target_length=8 , ) UpperCAmelCase__ : Optional[int] = train_dataset.dataset_kwargs assert "add_prefix_space" not in kwargs if tok_name != BART_TINY else "add_prefix_space" in kwargs assert len(_a ) == 1 if tok_name == BART_TINY else len(_a ) == 0
350
'''simple docstring''' import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin UpperCamelCase__ = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece @require_tokenizers class lowerCamelCase_ ( __a , unittest.TestCase ): lowerCAmelCase__ = XLMRobertaTokenizer lowerCAmelCase__ = XLMRobertaTokenizerFast lowerCAmelCase__ = True lowerCAmelCase__ = True def lowercase_ ( self : Dict ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase__ : Union[str, Any] = XLMRobertaTokenizer(_A , keep_accents=_A ) tokenizer.save_pretrained(self.tmpdirname ) def lowercase_ ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = '''<pad>''' UpperCAmelCase__ : Dict = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_A ) , _A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_A ) , _A ) def lowercase_ ( self : Any ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(_A ) , 1_002 ) def lowercase_ ( self : int ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 1_002 ) def lowercase_ ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : str = XLMRobertaTokenizer(_A , keep_accents=_A ) UpperCAmelCase__ : int = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(_A , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_A ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) UpperCAmelCase__ : Union[str, Any] = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( _A , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) UpperCAmelCase__ : Dict = tokenizer.convert_tokens_to_ids(_A ) self.assertListEqual( _A , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) UpperCAmelCase__ : Optional[int] = tokenizer.convert_ids_to_tokens(_A ) self.assertListEqual( _A , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) def lowercase_ ( self : str ): '''simple docstring''' if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return UpperCAmelCase__ : List[str] = (self.rust_tokenizer_class, '''hf-internal-testing/tiny-xlm-roberta''', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): UpperCAmelCase__ : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(_A , **_A ) UpperCAmelCase__ : Optional[int] = self.tokenizer_class.from_pretrained(_A , **_A ) UpperCAmelCase__ : List[str] = tempfile.mkdtemp() UpperCAmelCase__ : Any = tokenizer_r.save_pretrained(_A ) UpperCAmelCase__ : Tuple = tokenizer_p.save_pretrained(_A ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) UpperCAmelCase__ : Optional[int] = tuple(f for f in tokenizer_r_files if '''tokenizer.json''' not in f ) self.assertSequenceEqual(_A , _A ) # Checks everything loads correctly in the same way UpperCAmelCase__ : Any = tokenizer_r.from_pretrained(_A ) UpperCAmelCase__ : Dict = tokenizer_p.from_pretrained(_A ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_A , _A ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(_A ) # Save tokenizer rust, legacy_format=True UpperCAmelCase__ : Union[str, Any] = tempfile.mkdtemp() UpperCAmelCase__ : Union[str, Any] = tokenizer_r.save_pretrained(_A , legacy_format=_A ) UpperCAmelCase__ : List[str] = tokenizer_p.save_pretrained(_A ) # Checks it save with the same files self.assertSequenceEqual(_A , _A ) # Checks everything loads correctly in the same way UpperCAmelCase__ : List[str] = tokenizer_r.from_pretrained(_A ) UpperCAmelCase__ : List[str] = tokenizer_p.from_pretrained(_A ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_A , _A ) ) shutil.rmtree(_A ) # Save tokenizer rust, legacy_format=False UpperCAmelCase__ : Union[str, Any] = tempfile.mkdtemp() UpperCAmelCase__ : Dict = tokenizer_r.save_pretrained(_A , legacy_format=_A ) UpperCAmelCase__ : str = tokenizer_p.save_pretrained(_A ) # Checks it saved the tokenizer.json file self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way UpperCAmelCase__ : Union[str, Any] = tokenizer_r.from_pretrained(_A ) UpperCAmelCase__ : Optional[Any] = tokenizer_p.from_pretrained(_A ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_A , _A ) ) shutil.rmtree(_A ) @cached_property def lowercase_ ( self : Optional[Any] ): '''simple docstring''' return XLMRobertaTokenizer.from_pretrained('''xlm-roberta-base''' ) def lowercase_ ( self : Any ): '''simple docstring''' with tempfile.NamedTemporaryFile() as f: shutil.copyfile(_A , f.name ) UpperCAmelCase__ : int = XLMRobertaTokenizer(f.name , keep_accents=_A ) UpperCAmelCase__ : str = pickle.dumps(_A ) pickle.loads(_A ) def lowercase_ ( self : int ): '''simple docstring''' if not self.test_rust_tokenizer: return UpperCAmelCase__ : Optional[Any] = self.get_tokenizer() UpperCAmelCase__ : Union[str, Any] = self.get_rust_tokenizer() UpperCAmelCase__ : Dict = '''I was born in 92000, and this is falsé.''' UpperCAmelCase__ : Dict = tokenizer.tokenize(_A ) UpperCAmelCase__ : List[Any] = rust_tokenizer.tokenize(_A ) self.assertListEqual(_A , _A ) UpperCAmelCase__ : int = tokenizer.encode(_A , add_special_tokens=_A ) UpperCAmelCase__ : Optional[Any] = rust_tokenizer.encode(_A , add_special_tokens=_A ) self.assertListEqual(_A , _A ) UpperCAmelCase__ : Any = self.get_rust_tokenizer() UpperCAmelCase__ : List[Any] = tokenizer.encode(_A ) UpperCAmelCase__ : Union[str, Any] = rust_tokenizer.encode(_A ) self.assertListEqual(_A , _A ) @slow def lowercase_ ( self : str ): '''simple docstring''' UpperCAmelCase__ : str = '''Hello World!''' UpperCAmelCase__ : Tuple = [0, 35_378, 6_661, 38, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(_A , self.big_tokenizer.encode(_A ) ) @slow def lowercase_ ( self : Any ): '''simple docstring''' UpperCAmelCase__ : List[str] = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) UpperCAmelCase__ : Any = [ 0, 3_293, 83, 10, 4_552, 4_989, 7_986, 678, 10, 5_915, 111, 179_459, 124_850, 4, 6_044, 237, 12, 6, 5, 6, 4, 6_780, 705, 15, 1_388, 44, 378, 10_114, 711, 152, 20, 6, 5, 22_376, 642, 1_221, 15_190, 34_153, 450, 5_608, 959, 1_119, 57_702, 136, 186, 47, 1_098, 29_367, 47, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 6_044, 237, 6_284, 50_901, 528, 31, 90, 34, 927, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(_A , self.big_tokenizer.encode(_A ) ) @slow def lowercase_ ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : int = {'''input_ids''': [[0, 11_062, 82_772, 7, 15, 82_772, 538, 51_529, 237, 17_198, 1_290, 206, 9, 215_175, 1_314, 136, 17_198, 1_290, 206, 9, 56_359, 42, 122_009, 9, 16_466, 16, 87_344, 4_537, 9, 4_717, 78_381, 6, 159_958, 7, 15, 24_480, 618, 4, 527, 22_693, 5_428, 4, 2_777, 24_480, 9_874, 4, 43_523, 594, 4, 803, 18_392, 33_189, 18, 4, 43_523, 24_447, 12_399, 100, 24_955, 83_658, 9_626, 144_057, 15, 839, 22_335, 16, 136, 24_955, 83_658, 83_479, 15, 39_102, 724, 16, 678, 645, 2_789, 1_328, 4_589, 42, 122_009, 115_774, 23, 805, 1_328, 46_876, 7, 136, 53_894, 1_940, 42_227, 41_159, 17_721, 823, 425, 4, 27_512, 98_722, 206, 136, 5_531, 4_970, 919, 17_336, 5, 2], [0, 20_080, 618, 83, 82_775, 47, 479, 9, 1_517, 73, 53_894, 333, 80_581, 110_117, 18_811, 5_256, 1_295, 51, 152_526, 297, 7_986, 390, 124_416, 538, 35_431, 214, 98, 15_044, 25_737, 136, 7_108, 43_701, 23, 756, 135_355, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 581, 63_773, 119_455, 6, 147_797, 88_203, 7, 645, 70, 21, 3_285, 10_269, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_A , model_name='''xlm-roberta-base''' , revision='''d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3''' , )
299
0
def a ( snake_case__: int = 1_000 ): '''simple docstring''' lowercase_ = 2**power lowercase_ = 0 while n: lowercase_ , lowercase_ = r + n % 10, n // 10 return r if __name__ == "__main__": print(solution(int(str(input()).strip())))
30
def a ( snake_case__: list , snake_case__: list , snake_case__: int , snake_case__: int , snake_case__: int ): '''simple docstring''' if index == number_of_items: return 0 lowercase_ = 0 lowercase_ = 0 lowercase_ = knapsack(snake_case__ , snake_case__ , snake_case__ , snake_case__ , index + 1 ) if weights[index] <= max_weight: lowercase_ = values[index] + knapsack( snake_case__ , snake_case__ , snake_case__ , max_weight - weights[index] , index + 1 ) return max(snake_case__ , snake_case__ ) if __name__ == "__main__": import doctest doctest.testmod()
30
1
from __future__ import annotations class __lowerCamelCase : def __init__( self , lowerCamelCase=None ) -> Optional[int]: snake_case_ = data snake_case_ = None def __repr__( self ) -> List[str]: snake_case_ = [] snake_case_ = self while temp: string_rep.append(f'''{temp.data}''' ) snake_case_ = temp.next return "->".join(lowerCamelCase__ ) def UpperCamelCase( lowercase_ ) -> Optional[Any]: '''simple docstring''' if not elements_list: raise Exception("""The Elements List is empty""" ) snake_case_ = Node(elements_list[0] ) for i in range(1 , len(lowercase_ ) ): snake_case_ = Node(elements_list[i] ) snake_case_ = current.next return head def UpperCamelCase( lowercase_ ) -> Optional[int]: '''simple docstring''' if head_node is not None and isinstance(lowercase_ , lowercase_ ): print_reverse(head_node.next ) print(head_node.data ) def UpperCamelCase( ) -> Tuple: '''simple docstring''' from doctest import testmod testmod() snake_case_ = make_linked_list([14, 52, 14, 12, 43] ) print("""Linked List:""" ) print(lowercase_ ) print("""Elements in Reverse:""" ) print_reverse(lowercase_ ) if __name__ == "__main__": main()
356
import unittest from transformers import GPTNeoXJapaneseConfig, is_torch_available from transformers.models.gpt_neox_japanese.tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseModel class __lowerCamelCase : def __init__( self , lowerCamelCase , lowerCamelCase=13 , lowerCamelCase=7 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=99 , lowerCamelCase=32 , lowerCamelCase=5 , lowerCamelCase=4 , lowerCamelCase=4 , lowerCamelCase="gelu" , lowerCamelCase=0.0 , lowerCamelCase=0.1 , lowerCamelCase=True , lowerCamelCase=512 , lowerCamelCase=16 , lowerCamelCase=2 , lowerCamelCase=0.02 , lowerCamelCase=3 , lowerCamelCase=4 , lowerCamelCase=None , ) -> Union[str, Any]: snake_case_ = parent snake_case_ = batch_size snake_case_ = seq_length snake_case_ = is_training snake_case_ = use_input_mask snake_case_ = use_token_type_ids snake_case_ = use_labels snake_case_ = vocab_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_multiple_size snake_case_ = hidden_act snake_case_ = hidden_dropout snake_case_ = attention_dropout snake_case_ = weight_tying snake_case_ = max_position_embeddings snake_case_ = type_vocab_size snake_case_ = type_sequence_label_size snake_case_ = initializer_range snake_case_ = num_labels snake_case_ = num_choices snake_case_ = scope def lowerCAmelCase_ ( self ) -> str: snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ = None if self.use_input_mask: snake_case_ = random_attention_mask([self.batch_size, self.seq_length] ) snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case_ = self.get_config() return config, input_ids, input_mask, token_labels def lowerCAmelCase_ ( self ) -> Optional[int]: return GPTNeoXJapaneseConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_multiple_size=self.intermediate_multiple_size , hidden_act=self.hidden_act , hidden_dropout=self.hidden_dropout , attention_dropout=self.attention_dropout , weight_tying=self.weight_tying , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase , initializer_range=self.initializer_range , ) def lowerCAmelCase_ ( self ) -> int: snake_case_ , snake_case_ , snake_case_ , snake_case_ = self.prepare_config_and_inputs() snake_case_ = True return config, input_ids, input_mask, token_labels def lowerCAmelCase_ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ) -> Optional[int]: snake_case_ = GPTNeoXJapaneseModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() snake_case_ = model(lowerCamelCase , attention_mask=lowerCamelCase ) snake_case_ = model(lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase_ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ) -> Optional[int]: snake_case_ = True snake_case_ = GPTNeoXJapaneseModel(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() snake_case_ = model(lowerCamelCase , attention_mask=lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase_ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) -> Optional[Any]: snake_case_ = GPTNeoXJapaneseForCausalLM(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() snake_case_ = model(lowerCamelCase , attention_mask=lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCAmelCase_ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ) -> Tuple: snake_case_ = True snake_case_ = GPTNeoXJapaneseForCausalLM(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() # first forward pass snake_case_ = model(lowerCamelCase , attention_mask=lowerCamelCase , use_cache=lowerCamelCase ) snake_case_ = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids snake_case_ = ids_tensor((self.batch_size, 3) , config.vocab_size ) snake_case_ = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and snake_case_ = torch.cat([input_ids, next_tokens] , dim=-1 ) snake_case_ = torch.cat([input_mask, next_mask] , dim=-1 ) snake_case_ = model(lowerCamelCase , attention_mask=lowerCamelCase , output_hidden_states=lowerCamelCase ) snake_case_ = output_from_no_past["""hidden_states"""][0] snake_case_ = model( lowerCamelCase , attention_mask=lowerCamelCase , past_key_values=lowerCamelCase , output_hidden_states=lowerCamelCase , )["""hidden_states"""][0] # select random slice snake_case_ = ids_tensor((1,) , output_from_past.shape[-1] ).item() snake_case_ = output_from_no_past[:, -3:, random_slice_idx].detach() snake_case_ = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCamelCase , lowerCamelCase , atol=1e-3 ) ) def lowerCAmelCase_ ( self ) -> Union[str, Any]: snake_case_ = self.prepare_config_and_inputs() snake_case_ , snake_case_ , snake_case_ , snake_case_ = config_and_inputs snake_case_ = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __lowerCamelCase ( __snake_case , __snake_case , unittest.TestCase ): lowerCamelCase_ : Any = (GPTNeoXJapaneseModel, GPTNeoXJapaneseForCausalLM) if is_torch_available() else () lowerCamelCase_ : str = (GPTNeoXJapaneseForCausalLM,) if is_torch_available() else () lowerCamelCase_ : Any = ( {'feature-extraction': GPTNeoXJapaneseModel, 'text-generation': GPTNeoXJapaneseForCausalLM} if is_torch_available() else {} ) lowerCamelCase_ : Tuple = False lowerCamelCase_ : Dict = False lowerCamelCase_ : Tuple = False lowerCamelCase_ : Optional[int] = False def lowerCAmelCase_ ( self ) -> Tuple: snake_case_ = GPTNeoXJapaneseModelTester(self ) snake_case_ = ConfigTester(self , config_class=lowerCamelCase , hidden_size=37 ) def lowerCAmelCase_ ( self ) -> str: self.config_tester.run_common_tests() def lowerCAmelCase_ ( self ) -> str: snake_case_ , snake_case_ , snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def lowerCAmelCase_ ( self ) -> Tuple: snake_case_ , snake_case_ , snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def lowerCAmelCase_ ( self ) -> Tuple: # This regression test was failing with PyTorch < 1.3 snake_case_ , snake_case_ , snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_decoder() snake_case_ = None self.model_tester.create_and_check_model_as_decoder(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def lowerCAmelCase_ ( self ) -> Dict: snake_case_ , snake_case_ , snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def lowerCAmelCase_ ( self ) -> Dict: snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*lowerCamelCase ) @slow def lowerCAmelCase_ ( self ) -> Any: snake_case_ = """abeja/gpt-neox-japanese-2.7b""" snake_case_ = ["""データサイエンティストとは、""", """100年後に必要とされる会社は、""", """フルリモートの環境で働くために必要なことは、""", """国境の長いトンネルを抜けると""", """美味しい日本食といえば、"""] snake_case_ = [ """データサイエンティストとは、データを分析し、ビジネスに役立つ知見を導き出す専門家のことです。""", """100年後に必要とされる会社は、「人」が中心の会社です。""", """フルリモートの環境で働くために必要なことは、「自分の時間をコントロールする」ことです。""", """国境の長いトンネルを抜けると、そこは雪国だった。""", """美味しい日本食といえば、やっぱりお寿司ですよね。""", ] snake_case_ = GPTNeoXJapaneseTokenizer.from_pretrained(lowerCamelCase ) snake_case_ = GPTNeoXJapaneseForCausalLM.from_pretrained(lowerCamelCase ) snake_case_ = [] for prompt in prompts: snake_case_ = tokenizer(lowerCamelCase , return_tensors="""pt""" ).input_ids snake_case_ = model.generate(lowerCamelCase , max_length=50 ) snake_case_ = tokenizer.batch_decode(lowerCamelCase , skip_special_tokens=lowerCamelCase ) predicted_outputs += generated_string self.assertListEqual(lowerCamelCase , lowerCamelCase )
34
0
"""simple docstring""" from __future__ import annotations __UpperCAmelCase = [True] * 1_00_00_01 __UpperCAmelCase = 2 while i * i <= 1_00_00_00: if seive[i]: for j in range(i * i, 1_00_00_01, i): __UpperCAmelCase = False i += 1 def _snake_case ( lowercase__ : int ) -> bool: '''simple docstring''' return seive[n] def _snake_case ( lowercase__ : int ) -> bool: '''simple docstring''' return any(digit in """02468""" for digit in str(lowercase__ ) ) def _snake_case ( lowercase__ : int = 1_0_0_0_0_0_0 ) -> list[int]: '''simple docstring''' lowerCAmelCase_ :Tuple = [2] # result already includes the number 2. for num in range(3 , limit + 1 , 2 ): if is_prime(lowercase__ ) and not contains_an_even_digit(lowercase__ ): lowerCAmelCase_ :str = str(lowercase__ ) lowerCAmelCase_ :str = [int(str_num[j:] + str_num[:j] ) for j in range(len(lowercase__ ) )] if all(is_prime(lowercase__ ) for i in list_nums ): result.append(lowercase__ ) return result def _snake_case ( ) -> int: '''simple docstring''' return len(find_circular_primes() ) if __name__ == "__main__": print(F"""{len(find_circular_primes()) = }""")
84
import numpy as np def UpperCamelCase__( UpperCamelCase__ : np.array )->np.array: return (2 / (1 + np.exp(-2 * vector ))) - 1 if __name__ == "__main__": import doctest doctest.testmod()
193
0
'''simple docstring''' import torch from diffusers import KDPMaDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class UpperCamelCase_ (a__ ): """simple docstring""" _lowerCAmelCase = (KDPMaDiscreteScheduler,) _lowerCAmelCase = 1_0 def _a ( self : Tuple , **_lowerCamelCase : Dict ): """simple docstring""" A_ : Any = { '''num_train_timesteps''': 1100, '''beta_start''': 0.00_01, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', } config.update(**_lowerCamelCase ) return config def _a ( self : Tuple ): """simple docstring""" for timesteps in [10, 50, 100, 1000]: self.check_over_configs(num_train_timesteps=_lowerCamelCase ) def _a ( self : List[Any] ): """simple docstring""" for beta_start, beta_end in zip([0.0_00_01, 0.00_01, 0.0_01] , [0.00_02, 0.0_02, 0.02] ): self.check_over_configs(beta_start=_lowerCamelCase , beta_end=_lowerCamelCase ) def _a ( self : int ): """simple docstring""" for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=_lowerCamelCase ) def _a ( self : Union[str, Any] ): """simple docstring""" for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=_lowerCamelCase ) def _a ( self : Any ): """simple docstring""" A_ : Optional[int] = self.scheduler_classes[0] A_ : Optional[int] = self.get_scheduler_config(prediction_type='''v_prediction''' ) A_ : Any = scheduler_class(**_lowerCamelCase ) scheduler.set_timesteps(self.num_inference_steps ) A_ : Dict = self.dummy_model() A_ : int = self.dummy_sample_deter * scheduler.init_noise_sigma A_ : Optional[Any] = sample.to(_lowerCamelCase ) for i, t in enumerate(scheduler.timesteps ): A_ : List[Any] = scheduler.scale_model_input(_lowerCamelCase , _lowerCamelCase ) A_ : Tuple = model(_lowerCamelCase , _lowerCamelCase ) A_ : List[str] = scheduler.step(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : str = output.prev_sample A_ : int = torch.sum(torch.abs(_lowerCamelCase ) ) A_ : Any = torch.mean(torch.abs(_lowerCamelCase ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 4.6934E-07 ) < 1E-2 assert abs(result_mean.item() - 6.1112E-10 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 4.693_4286_5017_0972E-07 ) < 1E-2 assert abs(result_mean.item() - 0.00_02 ) < 1E-3 def _a ( self : List[Any] ): """simple docstring""" if torch_device == "mps": return A_ : Dict = self.scheduler_classes[0] A_ : List[Any] = self.get_scheduler_config() A_ : str = scheduler_class(**_lowerCamelCase ) scheduler.set_timesteps(self.num_inference_steps ) A_ : Any = self.dummy_model() A_ : Union[str, Any] = self.dummy_sample_deter * scheduler.init_noise_sigma A_ : Any = sample.to(_lowerCamelCase ) for i, t in enumerate(scheduler.timesteps ): A_ : str = scheduler.scale_model_input(_lowerCamelCase , _lowerCamelCase ) A_ : Dict = model(_lowerCamelCase , _lowerCamelCase ) A_ : Optional[Any] = scheduler.step(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : Any = output.prev_sample A_ : Optional[Any] = torch.sum(torch.abs(_lowerCamelCase ) ) A_ : Optional[Any] = torch.mean(torch.abs(_lowerCamelCase ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 20.41_25 ) < 1E-2 assert abs(result_mean.item() - 0.02_66 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 20.41_25 ) < 1E-2 assert abs(result_mean.item() - 0.02_66 ) < 1E-3 def _a ( self : List[Any] ): """simple docstring""" if torch_device == "mps": return A_ : Optional[Any] = self.scheduler_classes[0] A_ : Optional[int] = self.get_scheduler_config() A_ : List[str] = scheduler_class(**_lowerCamelCase ) scheduler.set_timesteps(self.num_inference_steps , device=_lowerCamelCase ) A_ : List[str] = self.dummy_model() A_ : List[Any] = self.dummy_sample_deter.to(_lowerCamelCase ) * scheduler.init_noise_sigma for t in scheduler.timesteps: A_ : Any = scheduler.scale_model_input(_lowerCamelCase , _lowerCamelCase ) A_ : Optional[Any] = model(_lowerCamelCase , _lowerCamelCase ) A_ : int = scheduler.step(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : str = output.prev_sample A_ : Optional[int] = torch.sum(torch.abs(_lowerCamelCase ) ) A_ : Dict = torch.mean(torch.abs(_lowerCamelCase ) ) if str(_lowerCamelCase ).startswith('''cpu''' ): # The following sum varies between 148 and 156 on mps. Why? assert abs(result_sum.item() - 20.41_25 ) < 1E-2 assert abs(result_mean.item() - 0.02_66 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 20.41_25 ) < 1E-2 assert abs(result_mean.item() - 0.02_66 ) < 1E-3
4
'''simple docstring''' import argparse from collections import OrderedDict from pathlib import Path import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision.transforms import functional as F from transformers import DetrImageProcessor, TableTransformerConfig, TableTransformerForObjectDetection from transformers.utils import logging logging.set_verbosity_info() snake_case__ = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) snake_case__ = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F'transformer.encoder.layers.{i}.self_attn.out_proj.weight', F'encoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (F'transformer.encoder.layers.{i}.self_attn.out_proj.bias', F'encoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append((F'transformer.encoder.layers.{i}.linear1.weight', F'encoder.layers.{i}.fc1.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.linear1.bias', F'encoder.layers.{i}.fc1.bias')) rename_keys.append((F'transformer.encoder.layers.{i}.linear2.weight', F'encoder.layers.{i}.fc2.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.linear2.bias', F'encoder.layers.{i}.fc2.bias')) rename_keys.append( (F'transformer.encoder.layers.{i}.norm1.weight', F'encoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((F'transformer.encoder.layers.{i}.norm1.bias', F'encoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append((F'transformer.encoder.layers.{i}.norm2.weight', F'encoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((F'transformer.encoder.layers.{i}.norm2.bias', F'encoder.layers.{i}.final_layer_norm.bias')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F'transformer.decoder.layers.{i}.self_attn.out_proj.weight', F'decoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.self_attn.out_proj.bias', F'decoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append( ( F'transformer.decoder.layers.{i}.multihead_attn.out_proj.weight', F'decoder.layers.{i}.encoder_attn.out_proj.weight', ) ) rename_keys.append( ( F'transformer.decoder.layers.{i}.multihead_attn.out_proj.bias', F'decoder.layers.{i}.encoder_attn.out_proj.bias', ) ) rename_keys.append((F'transformer.decoder.layers.{i}.linear1.weight', F'decoder.layers.{i}.fc1.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.linear1.bias', F'decoder.layers.{i}.fc1.bias')) rename_keys.append((F'transformer.decoder.layers.{i}.linear2.weight', F'decoder.layers.{i}.fc2.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.linear2.bias', F'decoder.layers.{i}.fc2.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.norm1.weight', F'decoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.norm1.bias', F'decoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append( (F'transformer.decoder.layers.{i}.norm2.weight', F'decoder.layers.{i}.encoder_attn_layer_norm.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.norm2.bias', F'decoder.layers.{i}.encoder_attn_layer_norm.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.norm3.weight', F'decoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((F'transformer.decoder.layers.{i}.norm3.bias', F'decoder.layers.{i}.final_layer_norm.bias')) # convolutional projection + query embeddings + layernorm of encoder + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("""input_proj.weight""", """input_projection.weight"""), ("""input_proj.bias""", """input_projection.bias"""), ("""query_embed.weight""", """query_position_embeddings.weight"""), ("""transformer.encoder.norm.weight""", """encoder.layernorm.weight"""), ("""transformer.encoder.norm.bias""", """encoder.layernorm.bias"""), ("""transformer.decoder.norm.weight""", """decoder.layernorm.weight"""), ("""transformer.decoder.norm.bias""", """decoder.layernorm.bias"""), ("""class_embed.weight""", """class_labels_classifier.weight"""), ("""class_embed.bias""", """class_labels_classifier.bias"""), ("""bbox_embed.layers.0.weight""", """bbox_predictor.layers.0.weight"""), ("""bbox_embed.layers.0.bias""", """bbox_predictor.layers.0.bias"""), ("""bbox_embed.layers.1.weight""", """bbox_predictor.layers.1.weight"""), ("""bbox_embed.layers.1.bias""", """bbox_predictor.layers.1.bias"""), ("""bbox_embed.layers.2.weight""", """bbox_predictor.layers.2.weight"""), ("""bbox_embed.layers.2.bias""", """bbox_predictor.layers.2.bias"""), ] ) def snake_case__ ( lowerCamelCase__ : List[Any] , lowerCamelCase__ : List[Any] , lowerCamelCase__ : List[Any] ) -> Optional[Any]: A_ : Tuple = state_dict.pop(lowerCamelCase__ ) A_ : Optional[Any] = val def snake_case__ ( lowerCamelCase__ : Dict ) -> Any: A_ : int = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: A_ : int = key.replace('''backbone.0.body''' , '''backbone.conv_encoder.model''' ) A_ : List[str] = value else: A_ : Optional[int] = value return new_state_dict def snake_case__ ( lowerCamelCase__ : Union[str, Any] ) -> Optional[Any]: A_ : Any = '''''' # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) A_ : Tuple = state_dict.pop(f'{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight' ) A_ : Dict = state_dict.pop(f'{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias' ) # next, add query, keys and values (in that order) to the state dict A_ : str = in_proj_weight[:2_5_6, :] A_ : Optional[Any] = in_proj_bias[:2_5_6] A_ : Dict = in_proj_weight[2_5_6:5_1_2, :] A_ : Tuple = in_proj_bias[2_5_6:5_1_2] A_ : Tuple = in_proj_weight[-2_5_6:, :] A_ : Optional[int] = in_proj_bias[-2_5_6:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention A_ : Union[str, Any] = state_dict.pop(f'{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight' ) A_ : Dict = state_dict.pop(f'{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias' ) # next, add query, keys and values (in that order) to the state dict A_ : List[str] = in_proj_weight[:2_5_6, :] A_ : int = in_proj_bias[:2_5_6] A_ : Any = in_proj_weight[2_5_6:5_1_2, :] A_ : List[str] = in_proj_bias[2_5_6:5_1_2] A_ : Union[str, Any] = in_proj_weight[-2_5_6:, :] A_ : Optional[Any] = in_proj_bias[-2_5_6:] # read in weights + bias of input projection layer of cross-attention A_ : Tuple = state_dict.pop( f'{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight' ) A_ : Optional[Any] = state_dict.pop(f'{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias' ) # next, add query, keys and values (in that order) of cross-attention to the state dict A_ : Dict = in_proj_weight_cross_attn[:2_5_6, :] A_ : Tuple = in_proj_bias_cross_attn[:2_5_6] A_ : int = in_proj_weight_cross_attn[2_5_6:5_1_2, :] A_ : List[str] = in_proj_bias_cross_attn[2_5_6:5_1_2] A_ : Any = in_proj_weight_cross_attn[-2_5_6:, :] A_ : Any = in_proj_bias_cross_attn[-2_5_6:] def snake_case__ ( lowerCamelCase__ : List[str] , lowerCamelCase__ : Tuple ) -> Dict: A_ ,A_ : int = image.size A_ : Tuple = max(lowerCamelCase__ , lowerCamelCase__ ) A_ : Optional[Any] = 8_0_0 if '''detection''' in checkpoint_url else 1_0_0_0 A_ : Union[str, Any] = target_max_size / current_max_size A_ : Any = image.resize((int(round(scale * width ) ), int(round(scale * height ) )) ) return resized_image def snake_case__ ( lowerCamelCase__ : Tuple ) -> str: A_ : Any = F.to_tensor(lowerCamelCase__ ) A_ : Optional[Any] = F.normalize(lowerCamelCase__ , mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ) return image @torch.no_grad() def snake_case__ ( lowerCamelCase__ : List[Any] , lowerCamelCase__ : int , lowerCamelCase__ : int ) -> str: logger.info('''Converting model...''' ) # load original state dict A_ : Tuple = torch.hub.load_state_dict_from_url(lowerCamelCase__ , map_location='''cpu''' ) # rename keys for src, dest in rename_keys: rename_key(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) A_ : str = rename_backbone_keys(lowerCamelCase__ ) # query, key and value matrices need special treatment read_in_q_k_v(lowerCamelCase__ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them A_ : List[Any] = '''model.''' for key in state_dict.copy().keys(): if not key.startswith('''class_labels_classifier''' ) and not key.startswith('''bbox_predictor''' ): A_ : List[Any] = state_dict.pop(lowerCamelCase__ ) A_ : str = val # create HuggingFace model and load state dict A_ : Union[str, Any] = TableTransformerConfig( backbone='''resnet18''' , mask_loss_coefficient=1 , dice_loss_coefficient=1 , ce_loss_coefficient=1 , bbox_loss_coefficient=5 , giou_loss_coefficient=2 , eos_coefficient=0.4 , class_cost=1 , bbox_cost=5 , giou_cost=2 , ) if "detection" in checkpoint_url: A_ : Dict = 1_5 A_ : Dict = 2 A_ : int = {0: '''table''', 1: '''table rotated'''} A_ : List[str] = idalabel A_ : Optional[int] = {v: k for k, v in idalabel.items()} else: A_ : Union[str, Any] = 1_2_5 A_ : Optional[Any] = 6 A_ : Optional[Any] = { 0: '''table''', 1: '''table column''', 2: '''table row''', 3: '''table column header''', 4: '''table projected row header''', 5: '''table spanning cell''', } A_ : int = idalabel A_ : Tuple = {v: k for k, v in idalabel.items()} A_ : Optional[Any] = DetrImageProcessor( format='''coco_detection''' , max_size=8_0_0 if '''detection''' in checkpoint_url else 1_0_0_0 ) A_ : int = TableTransformerForObjectDetection(lowerCamelCase__ ) model.load_state_dict(lowerCamelCase__ ) model.eval() # verify our conversion A_ : Optional[int] = '''example_pdf.png''' if '''detection''' in checkpoint_url else '''example_table.png''' A_ : Union[str, Any] = hf_hub_download(repo_id='''nielsr/example-pdf''' , repo_type='''dataset''' , filename=lowerCamelCase__ ) A_ : Tuple = Image.open(lowerCamelCase__ ).convert('''RGB''' ) A_ : int = normalize(resize(lowerCamelCase__ , lowerCamelCase__ ) ).unsqueeze(0 ) A_ : str = model(lowerCamelCase__ ) if "detection" in checkpoint_url: A_ : str = (1, 1_5, 3) A_ : int = torch.tensor( [[-6.7897, -16.9985, 6.7937], [-8.0186, -22.2192, 6.9677], [-7.3117, -21.0708, 7.4055]] ) A_ : Tuple = torch.tensor([[0.4867, 0.1767, 0.6732], [0.6718, 0.4479, 0.3830], [0.4716, 0.1760, 0.6364]] ) else: A_ : Optional[int] = (1, 1_2_5, 7) A_ : Dict = torch.tensor( [[-18.1430, -8.3214, 4.8274], [-18.4685, -7.1361, -4.2667], [-26.3693, -9.3429, -4.9962]] ) A_ : Any = torch.tensor([[0.4983, 0.5595, 0.9440], [0.4916, 0.6315, 0.5954], [0.6108, 0.8637, 0.1135]] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, :3, :3] , lowerCamelCase__ , atol=1e-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , lowerCamelCase__ , atol=1e-4 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(f'Saving PyTorch model and image processor to {pytorch_dump_folder_path}...' ) Path(lowerCamelCase__ ).mkdir(exist_ok=lowerCamelCase__ ) model.save_pretrained(lowerCamelCase__ ) image_processor.save_pretrained(lowerCamelCase__ ) if push_to_hub: # Push model to HF hub logger.info('''Pushing model to the hub...''' ) A_ : List[Any] = ( '''microsoft/table-transformer-detection''' if '''detection''' in checkpoint_url else '''microsoft/table-transformer-structure-recognition''' ) model.push_to_hub(lowerCamelCase__ ) image_processor.push_to_hub(lowerCamelCase__ ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser() parser.add_argument( """--checkpoint_url""", default="""https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", type=str, choices=[ """https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", """https://pubtables1m.blob.core.windows.net/model/pubtables1m_structure_detr_r18.pth""", ], help="""URL of the Table Transformer checkpoint you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) snake_case__ = parser.parse_args() convert_table_transformer_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
4
1
from __future__ import annotations from scipy.special import comb # type: ignore class lowercase : def __init__( self , _a ) -> Union[str, Any]: _A : Tuple = list_of_points # Degree determines the flexibility of the curve. # Degree = 1 will produce a straight line. _A : Dict = len(_a ) - 1 def a__ ( self , _a ) -> list[float]: assert 0 <= t <= 1, "Time t must be between 0 and 1." _A : list[float] = [] for i in range(len(self.list_of_points ) ): # basis function for each i output_values.append( comb(self.degree , _a ) * ((1 - t) ** (self.degree - i)) * (t**i) ) # the basis must sum up to 1 for it to produce a valid Bezier curve. assert round(sum(_a ) , 5 ) == 1 return output_values def a__ ( self , _a ) -> tuple[float, float]: assert 0 <= t <= 1, "Time t must be between 0 and 1." _A : Optional[int] = self.basis_function(_a ) _A : List[str] = 0.0 _A : Tuple = 0.0 for i in range(len(self.list_of_points ) ): # For all points, sum up the product of i-th basis function and i-th point. x += basis_function[i] * self.list_of_points[i][0] y += basis_function[i] * self.list_of_points[i][1] return (x, y) def a__ ( self , _a = 0.01 ) -> str: from matplotlib import pyplot as plt # type: ignore _A : list[float] = [] # x coordinates of points to plot _A : list[float] = [] # y coordinates of points to plot _A : str = 0.0 while t <= 1: _A : Optional[Any] = self.bezier_curve_function(_a ) to_plot_x.append(value[0] ) to_plot_y.append(value[1] ) t += step_size _A : Union[str, Any] = [i[0] for i in self.list_of_points] _A : str = [i[1] for i in self.list_of_points] plt.plot( _a , _a , color="""blue""" , label="""Curve of Degree """ + str(self.degree ) , ) plt.scatter(_a , _a , color="""red""" , label="""Control Points""" ) plt.legend() plt.show() if __name__ == "__main__": import doctest doctest.testmod() BezierCurve([(1, 2), (3, 5)]).plot_curve() # degree 1 BezierCurve([(0, 0), (5, 5), (5, 0)]).plot_curve() # degree 2 BezierCurve([(0, 0), (5, 5), (5, 0), (2.5, -2.5)]).plot_curve() # degree 3
26
import requests from bsa import BeautifulSoup def lowerCAmelCase_ ( snake_case_ = "AAPL" ): _A : str = f'''https://in.finance.yahoo.com/quote/{symbol}?s={symbol}''' _A : List[Any] = BeautifulSoup(requests.get(snake_case_ ).text,"""html.parser""" ) _A : Union[str, Any] = """My(6px) Pos(r) smartphone_Mt(6px)""" return soup.find("""div""",class_=class_ ).find("""span""" ).text if __name__ == "__main__": for symbol in "AAPL AMZN IBM GOOG MSFT ORCL".split(): print(f"""Current {symbol:<4} stock price is {stock_price(symbol):>8}""")
26
1
"""simple docstring""" import dataclasses import re from dataclasses import dataclass from functools import total_ordering from typing import Optional, Union __lowercase = re.compile(r'''^(?P<major>\d+)''' r'''\.(?P<minor>\d+)''' r'''\.(?P<patch>\d+)$''') @total_ordering @dataclass class _lowercase : """simple docstring""" lowercase__ = 42 lowercase__ = None lowercase__ = None lowercase__ = None lowercase__ = None def UpperCAmelCase_ ( self : Optional[Any] ) -> int: '''simple docstring''' __UpperCamelCase , __UpperCamelCase , __UpperCamelCase =_str_to_version_tuple(self.version_str ) def __repr__( self : Dict ) -> int: '''simple docstring''' return f"""{self.tuple[0]}.{self.tuple[1]}.{self.tuple[2]}""" @property def UpperCAmelCase_ ( self : Any ) -> int: '''simple docstring''' return self.major, self.minor, self.patch def UpperCAmelCase_ ( self : Optional[Any] , UpperCamelCase__ : Dict ) -> Union[str, Any]: '''simple docstring''' if isinstance(UpperCamelCase__ , UpperCamelCase__ ): return Version(UpperCamelCase__ ) elif isinstance(UpperCamelCase__ , UpperCamelCase__ ): return other raise TypeError(f"""{other} (type {type(UpperCamelCase__ )}) cannot be compared to version.""" ) def __eq__( self : Optional[Any] , UpperCamelCase__ : Union[str, Any] ) -> List[str]: '''simple docstring''' try: __UpperCamelCase =self._validate_operand(UpperCamelCase__ ) except (TypeError, ValueError): return False else: return self.tuple == other.tuple def __lt__( self : Union[str, Any] , UpperCamelCase__ : Union[str, Any] ) -> Optional[int]: '''simple docstring''' __UpperCamelCase =self._validate_operand(UpperCamelCase__ ) return self.tuple < other.tuple def __hash__( self : List[Any] ) -> List[str]: '''simple docstring''' return hash(_version_tuple_to_str(self.tuple ) ) @classmethod def UpperCAmelCase_ ( cls : Optional[Any] , UpperCamelCase__ : int ) -> Dict: '''simple docstring''' __UpperCamelCase ={f.name for f in dataclasses.fields(cls )} return cls(**{k: v for k, v in dic.items() if k in field_names} ) def UpperCAmelCase_ ( self : int ) -> str: '''simple docstring''' return self.version_str def lowerCAmelCase (__UpperCamelCase : List[Any] ): """simple docstring""" __UpperCamelCase =_VERSION_REG.match(__UpperCamelCase ) if not res: raise ValueError(F"""Invalid version '{version_str}'. Format should be x.y.z with {{x,y,z}} being digits.""" ) return tuple(int(__UpperCamelCase ) for v in [res.group('''major''' ), res.group('''minor''' ), res.group('''patch''' )] ) def lowerCAmelCase (__UpperCamelCase : int ): """simple docstring""" return ".".join(str(__UpperCamelCase ) for v in version_tuple )
352
"""simple docstring""" import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class _lowercase ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self : int ) -> int: '''simple docstring''' __UpperCamelCase ='''| <pad> <unk> <s> </s> a b c d e f g h i j k'''.split() __UpperCamelCase =dict(zip(UpperCamelCase__ , range(len(UpperCamelCase__ ) ) ) ) __UpperCamelCase ={ '''unk_token''': '''<unk>''', '''bos_token''': '''<s>''', '''eos_token''': '''</s>''', } __UpperCamelCase ={ '''feature_size''': 1, '''padding_value''': 0.0, '''sampling_rate''': 16000, '''return_attention_mask''': False, '''do_normalize''': True, } __UpperCamelCase =tempfile.mkdtemp() __UpperCamelCase =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) __UpperCamelCase =os.path.join(self.tmpdirname , UpperCamelCase__ ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(UpperCamelCase__ ) + '''\n''' ) with open(self.feature_extraction_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(UpperCamelCase__ ) + '''\n''' ) # load decoder from hub __UpperCamelCase ='''hf-internal-testing/ngram-beam-search-decoder''' def UpperCAmelCase_ ( self : Tuple , **UpperCamelCase__ : Tuple ) -> List[str]: '''simple docstring''' __UpperCamelCase =self.add_kwargs_tokens_map.copy() kwargs.update(UpperCamelCase__ ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def UpperCAmelCase_ ( self : Union[str, Any] , **UpperCamelCase__ : List[Any] ) -> Any: '''simple docstring''' return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def UpperCAmelCase_ ( self : List[Any] , **UpperCamelCase__ : Union[str, Any] ) -> str: '''simple docstring''' return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name , **UpperCamelCase__ ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> int: '''simple docstring''' shutil.rmtree(self.tmpdirname ) def UpperCAmelCase_ ( self : Tuple ) -> Optional[int]: '''simple docstring''' __UpperCamelCase =self.get_tokenizer() __UpperCamelCase =self.get_feature_extractor() __UpperCamelCase =self.get_decoder() __UpperCamelCase =WavaVecaProcessorWithLM(tokenizer=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , decoder=UpperCamelCase__ ) processor.save_pretrained(self.tmpdirname ) __UpperCamelCase =WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCamelCase__ ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , UpperCamelCase__ ) # decoder self.assertEqual(processor.decoder._alphabet.labels , decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set , decoder.model_container[decoder._model_key]._unigram_set , ) self.assertIsInstance(processor.decoder , UpperCamelCase__ ) def UpperCAmelCase_ ( self : Optional[int] ) -> str: '''simple docstring''' __UpperCamelCase =WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match __UpperCamelCase =WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname , alpha=5.0 , beta=3.0 , score_boundary=-7.0 , unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha , 5.0 ) self.assertEqual(processor.language_model.beta , 3.0 ) self.assertEqual(processor.language_model.score_boundary , -7.0 ) self.assertEqual(processor.language_model.unk_score_offset , 3 ) def UpperCAmelCase_ ( self : Optional[int] ) -> List[str]: '''simple docstring''' __UpperCamelCase =self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(['''xx'''] ) with self.assertRaisesRegex(UpperCamelCase__ , '''include''' ): WavaVecaProcessorWithLM( tokenizer=UpperCamelCase__ , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) def UpperCAmelCase_ ( self : List[Any] ) -> Dict: '''simple docstring''' __UpperCamelCase =self.get_feature_extractor() __UpperCamelCase =self.get_tokenizer() __UpperCamelCase =self.get_decoder() __UpperCamelCase =WavaVecaProcessorWithLM(tokenizer=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , decoder=UpperCamelCase__ ) __UpperCamelCase =floats_list((3, 1000) ) __UpperCamelCase =feature_extractor(UpperCamelCase__ , return_tensors='''np''' ) __UpperCamelCase =processor(UpperCamelCase__ , return_tensors='''np''' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCAmelCase_ ( self : List[str] ) -> Dict: '''simple docstring''' __UpperCamelCase =self.get_feature_extractor() __UpperCamelCase =self.get_tokenizer() __UpperCamelCase =self.get_decoder() __UpperCamelCase =WavaVecaProcessorWithLM(tokenizer=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , decoder=UpperCamelCase__ ) __UpperCamelCase ='''This is a test string''' __UpperCamelCase =processor(text=UpperCamelCase__ ) __UpperCamelCase =tokenizer(UpperCamelCase__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def UpperCAmelCase_ ( self : Union[str, Any] , UpperCamelCase__ : List[str]=(2, 10, 16) , UpperCamelCase__ : Union[str, Any]=77 ) -> int: '''simple docstring''' np.random.seed(UpperCamelCase__ ) return np.random.rand(*UpperCamelCase__ ) def UpperCAmelCase_ ( self : int ) -> Union[str, Any]: '''simple docstring''' __UpperCamelCase =self.get_feature_extractor() __UpperCamelCase =self.get_tokenizer() __UpperCamelCase =self.get_decoder() __UpperCamelCase =WavaVecaProcessorWithLM(tokenizer=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , decoder=UpperCamelCase__ ) __UpperCamelCase =self._get_dummy_logits(shape=(10, 16) , seed=13 ) __UpperCamelCase =processor.decode(UpperCamelCase__ ) __UpperCamelCase =decoder.decode_beams(UpperCamelCase__ )[0] self.assertEqual(decoded_decoder[0] , decoded_processor.text ) self.assertEqual('''</s> <s> </s>''' , decoded_processor.text ) self.assertEqual(decoded_decoder[-2] , decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] , decoded_processor.lm_score ) @parameterized.expand([[None], ['''fork'''], ['''spawn''']] ) def UpperCAmelCase_ ( self : Dict , UpperCamelCase__ : Optional[Any] ) -> List[str]: '''simple docstring''' __UpperCamelCase =self.get_feature_extractor() __UpperCamelCase =self.get_tokenizer() __UpperCamelCase =self.get_decoder() __UpperCamelCase =WavaVecaProcessorWithLM(tokenizer=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , decoder=UpperCamelCase__ ) __UpperCamelCase =self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: __UpperCamelCase =processor.batch_decode(UpperCamelCase__ ) else: with get_context(UpperCamelCase__ ).Pool() as pool: __UpperCamelCase =processor.batch_decode(UpperCamelCase__ , UpperCamelCase__ ) __UpperCamelCase =list(UpperCamelCase__ ) with get_context('''fork''' ).Pool() as p: __UpperCamelCase =decoder.decode_beams_batch(UpperCamelCase__ , UpperCamelCase__ ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase =[], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(UpperCamelCase__ , decoded_processor.text ) self.assertListEqual(['''<s> <s> </s>''', '''<s> <s> <s>'''] , decoded_processor.text ) self.assertListEqual(UpperCamelCase__ , decoded_processor.logit_score ) self.assertListEqual(UpperCamelCase__ , decoded_processor.lm_score ) def UpperCAmelCase_ ( self : Dict ) -> List[str]: '''simple docstring''' __UpperCamelCase =self.get_feature_extractor() __UpperCamelCase =self.get_tokenizer() __UpperCamelCase =self.get_decoder() __UpperCamelCase =WavaVecaProcessorWithLM(tokenizer=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , decoder=UpperCamelCase__ ) __UpperCamelCase =self._get_dummy_logits() __UpperCamelCase =15 __UpperCamelCase =-20.0 __UpperCamelCase =-4.0 __UpperCamelCase =processor.batch_decode( UpperCamelCase__ , beam_width=UpperCamelCase__ , beam_prune_logp=UpperCamelCase__ , token_min_logp=UpperCamelCase__ , ) __UpperCamelCase =decoded_processor_out.text __UpperCamelCase =list(UpperCamelCase__ ) with get_context('''fork''' ).Pool() as pool: __UpperCamelCase =decoder.decode_beams_batch( UpperCamelCase__ , UpperCamelCase__ , beam_width=UpperCamelCase__ , beam_prune_logp=UpperCamelCase__ , token_min_logp=UpperCamelCase__ , ) __UpperCamelCase =[d[0][0] for d in decoded_decoder_out] __UpperCamelCase =[d[0][2] for d in decoded_decoder_out] __UpperCamelCase =[d[0][3] for d in decoded_decoder_out] self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) self.assertListEqual(['''</s> <s> <s>''', '''<s> <s> <s>'''] , UpperCamelCase__ ) self.assertTrue(np.array_equal(UpperCamelCase__ , decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-20.0_54, -18.4_47] , UpperCamelCase__ , atol=1E-3 ) ) self.assertTrue(np.array_equal(UpperCamelCase__ , decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-15.5_54, -13.94_74] , UpperCamelCase__ , atol=1E-3 ) ) def UpperCAmelCase_ ( self : Any ) -> Optional[Any]: '''simple docstring''' __UpperCamelCase =self.get_feature_extractor() __UpperCamelCase =self.get_tokenizer() __UpperCamelCase =self.get_decoder() __UpperCamelCase =WavaVecaProcessorWithLM(tokenizer=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , decoder=UpperCamelCase__ ) __UpperCamelCase =self._get_dummy_logits() __UpperCamelCase =2.0 __UpperCamelCase =5.0 __UpperCamelCase =-20.0 __UpperCamelCase =True __UpperCamelCase =processor.batch_decode( UpperCamelCase__ , alpha=UpperCamelCase__ , beta=UpperCamelCase__ , unk_score_offset=UpperCamelCase__ , lm_score_boundary=UpperCamelCase__ , ) __UpperCamelCase =decoded_processor_out.text __UpperCamelCase =list(UpperCamelCase__ ) decoder.reset_params( alpha=UpperCamelCase__ , beta=UpperCamelCase__ , unk_score_offset=UpperCamelCase__ , lm_score_boundary=UpperCamelCase__ , ) with get_context('''fork''' ).Pool() as pool: __UpperCamelCase =decoder.decode_beams_batch( UpperCamelCase__ , UpperCamelCase__ , ) __UpperCamelCase =[d[0][0] for d in decoded_decoder_out] self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) self.assertListEqual(['''<s> </s> <s> </s> </s>''', '''</s> </s> <s> </s> </s>'''] , UpperCamelCase__ ) __UpperCamelCase =processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha , 2.0 ) self.assertEqual(lm_model.beta , 5.0 ) self.assertEqual(lm_model.unk_score_offset , -20.0 ) self.assertEqual(lm_model.score_boundary , UpperCamelCase__ ) def UpperCAmelCase_ ( self : Dict ) -> Optional[int]: '''simple docstring''' __UpperCamelCase =WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) __UpperCamelCase =processor.decoder.model_container[processor.decoder._model_key] __UpperCamelCase =Path(language_model._kenlm_model.path.decode('''utf-8''' ) ).parent.parent.absolute() __UpperCamelCase =os.listdir(UpperCamelCase__ ) __UpperCamelCase =['''alphabet.json''', '''language_model'''] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) def UpperCAmelCase_ ( self : List[str] ) -> Optional[Any]: '''simple docstring''' __UpperCamelCase =snapshot_download('''hf-internal-testing/processor_with_lm''' ) __UpperCamelCase =WavaVecaProcessorWithLM.from_pretrained(UpperCamelCase__ ) __UpperCamelCase =processor.decoder.model_container[processor.decoder._model_key] __UpperCamelCase =Path(language_model._kenlm_model.path.decode('''utf-8''' ) ).parent.parent.absolute() __UpperCamelCase =os.listdir(UpperCamelCase__ ) __UpperCamelCase =os.listdir(UpperCamelCase__ ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) def UpperCAmelCase_ ( self : Optional[int] ) -> List[str]: '''simple docstring''' __UpperCamelCase =WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) __UpperCamelCase =AutoProcessor.from_pretrained('''hf-internal-testing/processor_with_lm''' ) __UpperCamelCase =floats_list((3, 1000) ) __UpperCamelCase =processor_wavaveca(UpperCamelCase__ , return_tensors='''np''' ) __UpperCamelCase =processor_auto(UpperCamelCase__ , return_tensors='''np''' ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() , input_auto[key].sum() , delta=1E-2 ) __UpperCamelCase =self._get_dummy_logits() __UpperCamelCase =processor_wavaveca.batch_decode(UpperCamelCase__ ) __UpperCamelCase =processor_auto.batch_decode(UpperCamelCase__ ) self.assertListEqual(decoded_wavaveca.text , decoded_auto.text ) def UpperCAmelCase_ ( self : List[Any] ) -> int: '''simple docstring''' __UpperCamelCase =self.get_feature_extractor() __UpperCamelCase =self.get_tokenizer() __UpperCamelCase =self.get_decoder() __UpperCamelCase =WavaVecaProcessorWithLM(tokenizer=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , decoder=UpperCamelCase__ ) self.assertListEqual( processor.model_input_names , feature_extractor.model_input_names , msg='''`processor` and `feature_extractor` model input names do not match''' , ) @staticmethod def UpperCAmelCase_ ( UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Optional[Any] ) -> int: '''simple docstring''' __UpperCamelCase =[d[key] for d in offsets] return retrieved_list def UpperCAmelCase_ ( self : Dict ) -> List[str]: '''simple docstring''' __UpperCamelCase =WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) __UpperCamelCase =self._get_dummy_logits()[0] __UpperCamelCase =processor.decode(UpperCamelCase__ , output_word_offsets=UpperCamelCase__ ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('''text''' in outputs ) self.assertTrue('''word_offsets''' in outputs ) self.assertTrue(isinstance(UpperCamelCase__ , UpperCamelCase__ ) ) self.assertEqual(''' '''.join(self.get_from_offsets(outputs['''word_offsets'''] , '''word''' ) ) , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''word''' ) , ['''<s>''', '''<s>''', '''</s>'''] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''start_offset''' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''] , '''end_offset''' ) , [1, 3, 5] ) def UpperCAmelCase_ ( self : List[Any] ) -> Optional[int]: '''simple docstring''' __UpperCamelCase =WavaVecaProcessorWithLM.from_pretrained('''hf-internal-testing/processor_with_lm''' ) __UpperCamelCase =self._get_dummy_logits() __UpperCamelCase =processor.batch_decode(UpperCamelCase__ , output_word_offsets=UpperCamelCase__ ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue('''text''' in outputs ) self.assertTrue('''word_offsets''' in outputs ) self.assertTrue(isinstance(UpperCamelCase__ , UpperCamelCase__ ) ) self.assertListEqual( [''' '''.join(self.get_from_offsets(UpperCamelCase__ , '''word''' ) ) for o in outputs['''word_offsets''']] , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''word''' ) , ['''<s>''', '''<s>''', '''</s>'''] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''start_offset''' ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs['''word_offsets'''][0] , '''end_offset''' ) , [1, 3, 5] ) @slow @require_torch @require_torchaudio def UpperCAmelCase_ ( self : Optional[int] ) -> Tuple: '''simple docstring''' import torch __UpperCamelCase =load_dataset('''common_voice''' , '''en''' , split='''train''' , streaming=UpperCamelCase__ ) __UpperCamelCase =ds.cast_column('''audio''' , datasets.Audio(sampling_rate=16000 ) ) __UpperCamelCase =iter(UpperCamelCase__ ) __UpperCamelCase =next(UpperCamelCase__ ) __UpperCamelCase =AutoProcessor.from_pretrained('''patrickvonplaten/wav2vec2-base-100h-with-lm''' ) __UpperCamelCase =WavaVecaForCTC.from_pretrained('''patrickvonplaten/wav2vec2-base-100h-with-lm''' ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train __UpperCamelCase =processor(sample['''audio''']['''array'''] , return_tensors='''pt''' ).input_values with torch.no_grad(): __UpperCamelCase =model(UpperCamelCase__ ).logits.cpu().numpy() __UpperCamelCase =processor.decode(logits[0] , output_word_offsets=UpperCamelCase__ ) __UpperCamelCase =model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate __UpperCamelCase =[ { '''start_time''': d['''start_offset'''] * time_offset, '''end_time''': d['''end_offset'''] * time_offset, '''word''': d['''word'''], } for d in output['''word_offsets'''] ] __UpperCamelCase ='''WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL''' # output words self.assertEqual(''' '''.join(self.get_from_offsets(UpperCamelCase__ , '''word''' ) ) , UpperCamelCase__ ) self.assertEqual(''' '''.join(self.get_from_offsets(UpperCamelCase__ , '''word''' ) ) , output.text ) # output times __UpperCamelCase =torch.tensor(self.get_from_offsets(UpperCamelCase__ , '''start_time''' ) ) __UpperCamelCase =torch.tensor(self.get_from_offsets(UpperCamelCase__ , '''end_time''' ) ) # fmt: off __UpperCamelCase =torch.tensor([1.41_99, 1.65_99, 2.25_99, 3.0, 3.24, 3.59_99, 3.79_99, 4.09_99, 4.26, 4.94, 5.28, 5.65_99, 5.78, 5.94, 6.32, 6.53_99, 6.65_99] ) __UpperCamelCase =torch.tensor([1.53_99, 1.89_99, 2.9, 3.16, 3.53_99, 3.72, 4.01_99, 4.17_99, 4.76, 5.15_99, 5.55_99, 5.69_99, 5.86, 6.19_99, 6.38, 6.61_99, 6.94] ) # fmt: on self.assertTrue(torch.allclose(UpperCamelCase__ , UpperCamelCase__ , atol=0.01 ) ) self.assertTrue(torch.allclose(UpperCamelCase__ , UpperCamelCase__ , atol=0.01 ) )
85
0
from math import factorial def _lowerCAmelCase ( __lowerCAmelCase = 20 ) -> int: """simple docstring""" snake_case__ : str = 2 * n # middle entry of odd rows starting at row 3 is the solution for n = 1, # 2, 3,... snake_case__ : Dict = n // 2 return int(factorial(__A ) / (factorial(__A ) * factorial(n - k )) ) if __name__ == "__main__": import sys if len(sys.argv) == 1: print(solution(20)) else: try: A__ = int(sys.argv[1]) print(solution(n)) except ValueError: print('''Invalid entry - please enter a number.''')
230
import jax.numpy as jnp from ...utils import logging from ..ta.modeling_flax_ta import FlaxTaEncoderModel, FlaxTaForConditionalGeneration, FlaxTaModel from .configuration_mta import MTaConfig UpperCAmelCase_ : Optional[Any] = logging.get_logger(__name__) UpperCAmelCase_ : str = 'T5Config' def SCREAMING_SNAKE_CASE_ ( __A : jnp.array , __A : int , __A : int ) -> jnp.ndarray: """simple docstring""" a_ : Dict = jnp.zeros_like(__A ) a_ : Dict = shifted_input_ids.at[:, 1:].set(input_ids[:, :-1] ) a_ : str = shifted_input_ids.at[:, 0].set(__A ) a_ : int = jnp.where(shifted_input_ids == -1_00 , __A , __A ) return shifted_input_ids class SCREAMING_SNAKE_CASE__ ( lowercase__ ): snake_case__ : str = '''mt5''' snake_case__ : List[Any] = MTaConfig class SCREAMING_SNAKE_CASE__ ( lowercase__ ): snake_case__ : str = '''mt5''' snake_case__ : List[str] = MTaConfig class SCREAMING_SNAKE_CASE__ ( lowercase__ ): snake_case__ : Any = '''mt5''' snake_case__ : Union[str, Any] = MTaConfig
32
0
'''simple docstring''' import doctest from collections import deque import numpy as np class lowercase : """simple docstring""" def __init__( self ): '''simple docstring''' UpperCamelCase__ :Dict = [2, 1, 2, -1] UpperCamelCase__ :List[Any] = [1, 2, 3, 4] def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Union[str, Any] = len(self.first_signal ) UpperCamelCase__ :int = len(self.second_signal ) UpperCamelCase__ :int = max(UpperCamelCase_ , UpperCamelCase_ ) # create a zero matrix of max_length x max_length UpperCamelCase__ :Optional[int] = [[0] * max_length for i in range(UpperCamelCase_ )] # fills the smaller signal with zeros to make both signals of same length if length_first_signal < length_second_signal: self.first_signal += [0] * (max_length - length_first_signal) elif length_first_signal > length_second_signal: self.second_signal += [0] * (max_length - length_second_signal) for i in range(UpperCamelCase_ ): UpperCamelCase__ :Optional[Any] = deque(self.second_signal ) rotated_signal.rotate(UpperCamelCase_ ) for j, item in enumerate(UpperCamelCase_ ): matrix[i][j] += item # multiply the matrix with the first signal UpperCamelCase__ :Dict = np.matmul(np.transpose(UpperCamelCase_ ) , np.transpose(self.first_signal ) ) # rounding-off to two decimal places return [round(UpperCamelCase_ , 2 ) for i in final_signal] if __name__ == "__main__": doctest.testmod()
355
'''simple docstring''' import logging from transformers.configuration_utils import PretrainedConfig __snake_case = logging.getLogger(__name__) class lowercase ( A__ ): """simple docstring""" _a = 'masked_bert' def __init__( self , UpperCamelCase_=30522 , UpperCamelCase_=768 , UpperCamelCase_=12 , UpperCamelCase_=12 , UpperCamelCase_=3072 , UpperCamelCase_="gelu" , UpperCamelCase_=0.1 , UpperCamelCase_=0.1 , UpperCamelCase_=512 , UpperCamelCase_=2 , UpperCamelCase_=0.02 , UpperCamelCase_=1e-12 , UpperCamelCase_=0 , UpperCamelCase_="topK" , UpperCamelCase_="constant" , UpperCamelCase_=0.0 , **UpperCamelCase_ , ): '''simple docstring''' super().__init__(pad_token_id=UpperCamelCase_ , **UpperCamelCase_ ) UpperCamelCase__ :Tuple = vocab_size UpperCamelCase__ :Dict = hidden_size UpperCamelCase__ :Optional[int] = num_hidden_layers UpperCamelCase__ :Any = num_attention_heads UpperCamelCase__ :Optional[Any] = hidden_act UpperCamelCase__ :str = intermediate_size UpperCamelCase__ :List[Any] = hidden_dropout_prob UpperCamelCase__ :List[str] = attention_probs_dropout_prob UpperCamelCase__ :List[Any] = max_position_embeddings UpperCamelCase__ :int = type_vocab_size UpperCamelCase__ :List[Any] = initializer_range UpperCamelCase__ :Any = layer_norm_eps UpperCamelCase__ :str = pruning_method UpperCamelCase__ :int = mask_init UpperCamelCase__ :Optional[Any] = mask_scale
219
0
import argparse import json from dataclasses import dataclass, field from functools import partial from pathlib import Path from typing import List import timm import torch import torch.nn as nn from huggingface_hub import hf_hub_download from torch import Tensor from transformers import AutoImageProcessor, ResNetConfig, ResNetForImageClassification from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase__ = logging.get_logger() @dataclass class a__ : """simple docstring""" __lowerCamelCase = 42 __lowerCamelCase = field(default_factory=A__ ) __lowerCamelCase = field(default_factory=A__ ) def UpperCamelCase ( self , lowercase , lowercase , lowercase ) -> List[Any]: '''simple docstring''' A__ = len(list(m.modules() ) ) == 1 or isinstance(__A , nn.Convad ) or isinstance(__A , nn.BatchNormad ) if has_not_submodules: self.traced.append(__A ) def __call__( self , lowercase ) -> Union[str, Any]: '''simple docstring''' for m in self.module.modules(): self.handles.append(m.register_forward_hook(self._forward_hook ) ) self.module(__A ) [x.remove() for x in self.handles] return self @property def UpperCamelCase ( self ) -> List[str]: '''simple docstring''' return list(filter(lambda lowercase : len(list(x.state_dict().keys() ) ) > 0 , self.traced ) ) @dataclass class a__ : """simple docstring""" __lowerCamelCase = 42 __lowerCamelCase = 42 __lowerCamelCase = 0 __lowerCamelCase = field(default_factory=A__ ) __lowerCamelCase = field(default_factory=A__ ) def __call__( self , lowercase ) -> List[str]: '''simple docstring''' A__ = Tracker(self.dest )(__A ).parametrized A__ = Tracker(self.src )(__A ).parametrized A__ = list(filter(lambda lowercase : type(__A ) not in self.src_skip , __A ) ) A__ = list(filter(lambda lowercase : type(__A ) not in self.dest_skip , __A ) ) if len(__A ) != len(__A ): raise Exception( F'Numbers of operations are different. Source module has {len(__A )} operations while' F' destination module has {len(__A )}.' ) for dest_m, src_m in zip(__A , __A ): dest_m.load_state_dict(src_m.state_dict() ) if self.verbose == 1: print(F'Transfered from={src_m} to={dest_m}' ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: str , SCREAMING_SNAKE_CASE_: ResNetConfig , SCREAMING_SNAKE_CASE_: Path , SCREAMING_SNAKE_CASE_: bool = True ) -> Union[str, Any]: '''simple docstring''' print(F'Converting {name}...' ) with torch.no_grad(): A__ = timm.create_model(lowercase__ , pretrained=lowercase__ ).eval() A__ = ResNetForImageClassification(lowercase__ ).eval() A__ = ModuleTransfer(src=lowercase__ , dest=lowercase__ ) A__ = torch.randn((1, 3, 2_2_4, 2_2_4) ) module_transfer(lowercase__ ) assert torch.allclose(from_model(lowercase__ ) , our_model(lowercase__ ).logits ), "The model logits don't match the original one." A__ = F'resnet{"-".join(name.split("resnet" ) )}' print(lowercase__ ) if push_to_hub: our_model.push_to_hub( repo_path_or_name=save_directory / checkpoint_name , commit_message="Add model" , use_temp_dir=lowercase__ , ) # we can use the convnext one A__ = AutoImageProcessor.from_pretrained("facebook/convnext-base-224-22k-1k" ) image_processor.push_to_hub( repo_path_or_name=save_directory / checkpoint_name , commit_message="Add image processor" , use_temp_dir=lowercase__ , ) print(F'Pushed {checkpoint_name}' ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: Path , SCREAMING_SNAKE_CASE_: str = None , SCREAMING_SNAKE_CASE_: bool = True ) -> List[Any]: '''simple docstring''' A__ = """imagenet-1k-id2label.json""" A__ = 1_0_0_0 A__ = (1, num_labels) A__ = """huggingface/label-files""" A__ = num_labels A__ = json.load(open(hf_hub_download(lowercase__ , lowercase__ , repo_type="dataset" ) , "r" ) ) A__ = {int(lowercase__ ): v for k, v in idalabel.items()} A__ = idalabel A__ = {v: k for k, v in idalabel.items()} A__ = partial(lowercase__ , num_labels=lowercase__ , idalabel=lowercase__ , labelaid=lowercase__ ) A__ = { """resnet18""": ImageNetPreTrainedConfig( depths=[2, 2, 2, 2] , hidden_sizes=[6_4, 1_2_8, 2_5_6, 5_1_2] , layer_type="basic" ), """resnet26""": ImageNetPreTrainedConfig( depths=[2, 2, 2, 2] , hidden_sizes=[2_5_6, 5_1_2, 1_0_2_4, 2_0_4_8] , layer_type="bottleneck" ), """resnet34""": ImageNetPreTrainedConfig( depths=[3, 4, 6, 3] , hidden_sizes=[6_4, 1_2_8, 2_5_6, 5_1_2] , layer_type="basic" ), """resnet50""": ImageNetPreTrainedConfig( depths=[3, 4, 6, 3] , hidden_sizes=[2_5_6, 5_1_2, 1_0_2_4, 2_0_4_8] , layer_type="bottleneck" ), """resnet101""": ImageNetPreTrainedConfig( depths=[3, 4, 2_3, 3] , hidden_sizes=[2_5_6, 5_1_2, 1_0_2_4, 2_0_4_8] , layer_type="bottleneck" ), """resnet152""": ImageNetPreTrainedConfig( depths=[3, 8, 3_6, 3] , hidden_sizes=[2_5_6, 5_1_2, 1_0_2_4, 2_0_4_8] , layer_type="bottleneck" ), } if model_name: convert_weight_and_push(lowercase__ , names_to_config[model_name] , lowercase__ , lowercase__ ) else: for model_name, config in names_to_config.items(): convert_weight_and_push(lowercase__ , lowercase__ , lowercase__ , lowercase__ ) return config, expected_shape if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default=None, type=str, help=( """The name of the model you wish to convert, it must be one of the supported resnet* architecture,""" """ currently: resnet18,26,34,50,101,152. If `None`, all of them will the converted.""" ), ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=Path, required=True, help="""Path to the output PyTorch model directory.""", ) parser.add_argument( """--push_to_hub""", default=True, type=bool, required=False, help="""If True, push model and image processor to the hub.""", ) lowerCAmelCase__ = parser.parse_args() lowerCAmelCase__ = args.pytorch_dump_folder_path pytorch_dump_folder_path.mkdir(exist_ok=True, parents=True) convert_weights_and_push(pytorch_dump_folder_path, args.model_name, args.push_to_hub)
68
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import requests import torch from PIL import Image from transformers import GLPNConfig, GLPNForDepthEstimation, GLPNImageProcessor from transformers.utils import logging logging.set_verbosity_info() __UpperCAmelCase = logging.get_logger(__name__) def _snake_case ( lowercase__ : Optional[Any] ) -> str: '''simple docstring''' lowerCAmelCase_ :str = OrderedDict() for key, value in state_dict.items(): if key.startswith("""module.encoder""" ): lowerCAmelCase_ :Union[str, Any] = key.replace("""module.encoder""" , """glpn.encoder""" ) if key.startswith("""module.decoder""" ): lowerCAmelCase_ :Any = key.replace("""module.decoder""" , """decoder.stages""" ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 lowerCAmelCase_ :List[str] = key[key.find("""patch_embed""" ) + len("""patch_embed""" )] lowerCAmelCase_ :Tuple = key.replace(f"""patch_embed{idx}""" , f"""patch_embeddings.{int(lowercase__ )-1}""" ) if "norm" in key: lowerCAmelCase_ :Dict = key.replace("""norm""" , """layer_norm""" ) if "glpn.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 lowerCAmelCase_ :str = key[key.find("""glpn.encoder.layer_norm""" ) + len("""glpn.encoder.layer_norm""" )] lowerCAmelCase_ :str = key.replace(f"""layer_norm{idx}""" , f"""layer_norm.{int(lowercase__ )-1}""" ) if "layer_norm1" in key: lowerCAmelCase_ :Optional[Any] = key.replace("""layer_norm1""" , """layer_norm_1""" ) if "layer_norm2" in key: lowerCAmelCase_ :str = key.replace("""layer_norm2""" , """layer_norm_2""" ) if "block" in key: # replace for example block1 by block.0 lowerCAmelCase_ :List[str] = key[key.find("""block""" ) + len("""block""" )] lowerCAmelCase_ :int = key.replace(f"""block{idx}""" , f"""block.{int(lowercase__ )-1}""" ) if "attn.q" in key: lowerCAmelCase_ :Tuple = key.replace("""attn.q""" , """attention.self.query""" ) if "attn.proj" in key: lowerCAmelCase_ :Optional[int] = key.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in key: lowerCAmelCase_ :str = key.replace("""attn""" , """attention.self""" ) if "fc1" in key: lowerCAmelCase_ :List[Any] = key.replace("""fc1""" , """dense1""" ) if "fc2" in key: lowerCAmelCase_ :Optional[Any] = key.replace("""fc2""" , """dense2""" ) if "linear_pred" in key: lowerCAmelCase_ :List[str] = key.replace("""linear_pred""" , """classifier""" ) if "linear_fuse" in key: lowerCAmelCase_ :str = key.replace("""linear_fuse.conv""" , """linear_fuse""" ) lowerCAmelCase_ :Any = key.replace("""linear_fuse.bn""" , """batch_norm""" ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 lowerCAmelCase_ :str = key[key.find("""linear_c""" ) + len("""linear_c""" )] lowerCAmelCase_ :Optional[int] = key.replace(f"""linear_c{idx}""" , f"""linear_c.{int(lowercase__ )-1}""" ) if "bot_conv" in key: lowerCAmelCase_ :Union[str, Any] = key.replace("""bot_conv""" , """0.convolution""" ) if "skip_conv1" in key: lowerCAmelCase_ :int = key.replace("""skip_conv1""" , """1.convolution""" ) if "skip_conv2" in key: lowerCAmelCase_ :str = key.replace("""skip_conv2""" , """2.convolution""" ) if "fusion1" in key: lowerCAmelCase_ :Any = key.replace("""fusion1""" , """1.fusion""" ) if "fusion2" in key: lowerCAmelCase_ :List[str] = key.replace("""fusion2""" , """2.fusion""" ) if "fusion3" in key: lowerCAmelCase_ :Dict = key.replace("""fusion3""" , """3.fusion""" ) if "fusion" in key and "conv" in key: lowerCAmelCase_ :Any = key.replace("""conv""" , """convolutional_layer""" ) if key.startswith("""module.last_layer_depth""" ): lowerCAmelCase_ :Tuple = key.replace("""module.last_layer_depth""" , """head.head""" ) lowerCAmelCase_ :List[Any] = value return new_state_dict def _snake_case ( lowercase__ : str , lowercase__ : int ) -> str: '''simple docstring''' for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) lowerCAmelCase_ :Tuple = state_dict.pop(f"""glpn.encoder.block.{i}.{j}.attention.self.kv.weight""" ) lowerCAmelCase_ :Tuple = state_dict.pop(f"""glpn.encoder.block.{i}.{j}.attention.self.kv.bias""" ) # next, add keys and values (in that order) to the state dict lowerCAmelCase_ :Optional[Any] = kv_weight[ : config.hidden_sizes[i], : ] lowerCAmelCase_ :Union[str, Any] = kv_bias[: config.hidden_sizes[i]] lowerCAmelCase_ :List[Any] = kv_weight[ config.hidden_sizes[i] :, : ] lowerCAmelCase_ :int = kv_bias[config.hidden_sizes[i] :] def _snake_case ( ) -> Any: '''simple docstring''' lowerCAmelCase_ :int = """http://images.cocodataset.org/val2017/000000039769.jpg""" lowerCAmelCase_ :Optional[Any] = Image.open(requests.get(lowercase__ , stream=lowercase__ ).raw ) return image @torch.no_grad() def _snake_case ( lowercase__ : List[Any] , lowercase__ : str , lowercase__ : Dict=False , lowercase__ : List[Any]=None ) -> int: '''simple docstring''' lowerCAmelCase_ :int = GLPNConfig(hidden_sizes=[6_4, 1_2_8, 3_2_0, 5_1_2] , decoder_hidden_size=6_4 , depths=[3, 8, 2_7, 3] ) # load image processor (only resize + rescale) lowerCAmelCase_ :Union[str, Any] = GLPNImageProcessor() # prepare image lowerCAmelCase_ :List[Any] = prepare_img() lowerCAmelCase_ :int = image_processor(images=lowercase__ , return_tensors="""pt""" ).pixel_values logger.info("""Converting model...""" ) # load original state dict lowerCAmelCase_ :Tuple = torch.load(lowercase__ , map_location=torch.device("""cpu""" ) ) # rename keys lowerCAmelCase_ :Union[str, Any] = rename_keys(lowercase__ ) # key and value matrices need special treatment read_in_k_v(lowercase__ , lowercase__ ) # create HuggingFace model and load state dict lowerCAmelCase_ :List[Any] = GLPNForDepthEstimation(lowercase__ ) model.load_state_dict(lowercase__ ) model.eval() # forward pass lowerCAmelCase_ :Dict = model(lowercase__ ) lowerCAmelCase_ :Tuple = outputs.predicted_depth # verify output if model_name is not None: if "nyu" in model_name: lowerCAmelCase_ :Optional[Any] = torch.tensor( [[4.4147, 4.0873, 4.0673], [3.7890, 3.2881, 3.1525], [3.7674, 3.5423, 3.4913]] ) elif "kitti" in model_name: lowerCAmelCase_ :Any = torch.tensor( [[3.4291, 2.7865, 2.5151], [3.2841, 2.7021, 2.3502], [3.1147, 2.4625, 2.2481]] ) else: raise ValueError(f"""Unknown model name: {model_name}""" ) lowerCAmelCase_ :Union[str, Any] = torch.Size([1, 4_8_0, 6_4_0] ) assert predicted_depth.shape == expected_shape assert torch.allclose(predicted_depth[0, :3, :3] , lowercase__ , atol=1E-4 ) print("""Looks ok!""" ) # finally, push to hub if required if push_to_hub: logger.info("""Pushing model and image processor to the hub...""" ) model.push_to_hub( repo_path_or_name=Path(lowercase__ , lowercase__ ) , organization="""nielsr""" , commit_message="""Add model""" , use_temp_dir=lowercase__ , ) image_processor.push_to_hub( repo_path_or_name=Path(lowercase__ , lowercase__ ) , organization="""nielsr""" , commit_message="""Add image processor""" , use_temp_dir=lowercase__ , ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() parser.add_argument( '--checkpoint_path', default=None, type=str, help='Path to the original PyTorch checkpoint (.pth file).', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether to upload the model to the HuggingFace hub.' ) parser.add_argument( '--model_name', default='glpn-kitti', type=str, help='Name of the model in case you\'re pushing to the hub.', ) __UpperCAmelCase = parser.parse_args() convert_glpn_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
84
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { 'unc-nlp/lxmert-base-uncased': 'https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json', } class UpperCamelCase ( snake_case_ ): UpperCamelCase : int = '''lxmert''' UpperCamelCase : Tuple = {} def __init__( self : Dict , UpperCAmelCase__ : Union[str, Any]=30522 , UpperCAmelCase__ : str=768 , UpperCAmelCase__ : Any=12 , UpperCAmelCase__ : Dict=9500 , UpperCAmelCase__ : Tuple=1600 , UpperCAmelCase__ : int=400 , UpperCAmelCase__ : List[str]=3072 , UpperCAmelCase__ : Optional[Any]="gelu" , UpperCAmelCase__ : Any=0.1 , UpperCAmelCase__ : List[Any]=0.1 , UpperCAmelCase__ : int=512 , UpperCAmelCase__ : str=2 , UpperCAmelCase__ : Union[str, Any]=0.0_2 , UpperCAmelCase__ : Optional[int]=1E-12 , UpperCAmelCase__ : int=9 , UpperCAmelCase__ : Union[str, Any]=5 , UpperCAmelCase__ : Optional[Any]=5 , UpperCAmelCase__ : Tuple=2048 , UpperCAmelCase__ : List[Any]=4 , UpperCAmelCase__ : Union[str, Any]=6.6_7 , UpperCAmelCase__ : Any=True , UpperCAmelCase__ : str=True , UpperCAmelCase__ : str=True , UpperCAmelCase__ : Tuple=True , UpperCAmelCase__ : List[Any]=True , UpperCAmelCase__ : str=True , UpperCAmelCase__ : str=True , **UpperCAmelCase__ : List[Any] , ) -> Optional[int]: _a : Optional[int] = vocab_size _a : List[str] = hidden_size _a : Any = num_attention_heads _a : str = hidden_act _a : Union[str, Any] = intermediate_size _a : int = hidden_dropout_prob _a : List[str] = attention_probs_dropout_prob _a : List[str] = max_position_embeddings _a : Dict = type_vocab_size _a : int = initializer_range _a : Union[str, Any] = layer_norm_eps _a : Optional[int] = num_qa_labels _a : int = num_object_labels _a : List[Any] = num_attr_labels _a : List[Any] = l_layers _a : Tuple = x_layers _a : Union[str, Any] = r_layers _a : Dict = visual_feat_dim _a : Optional[int] = visual_pos_dim _a : Dict = visual_loss_normalizer _a : Any = task_matched _a : int = task_mask_lm _a : Any = task_obj_predict _a : int = task_qa _a : Union[str, Any] = visual_obj_loss _a : str = visual_attr_loss _a : List[Any] = visual_feat_loss _a : Optional[int] = {"""vision""": r_layers, """cross_encoder""": x_layers, """language""": l_layers} super().__init__(**UpperCAmelCase__ )
324
"""simple docstring""" import os from huggingface_hub.constants import HUGGINGFACE_HUB_CACHE, hf_cache_home _snake_case = HUGGINGFACE_HUB_CACHE _snake_case = 'config.json' _snake_case = 'diffusion_pytorch_model.bin' _snake_case = 'diffusion_flax_model.msgpack' _snake_case = 'model.onnx' _snake_case = 'diffusion_pytorch_model.safetensors' _snake_case = 'weights.pb' _snake_case = 'https://huggingface.co' _snake_case = default_cache_path _snake_case = 'diffusers_modules' _snake_case = os.getenv('HF_MODULES_CACHE', os.path.join(hf_cache_home, 'modules')) _snake_case = ['fp16', 'non-ema'] _snake_case = '.self_attn'
324
1
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TextClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. lowercase_ = {"""LayoutLMv2Config""", """LayoutLMv3Config"""} @is_pipeline_test class a_ ( unittest.TestCase ): '''simple docstring''' UpperCamelCase = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING UpperCamelCase = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: UpperCamelCase = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: UpperCamelCase = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } @require_torch def snake_case_( self ) -> Union[str, Any]: _SCREAMING_SNAKE_CASE = pipeline( task="""text-classification""" , model="""hf-internal-testing/tiny-random-distilbert""" , framework="""pt""" ) _SCREAMING_SNAKE_CASE = text_classifier("""This is great !""" ) self.assertEqual(nested_simplify(A ) , [{"""label""": """LABEL_0""", """score""": 0.504}] ) _SCREAMING_SNAKE_CASE = text_classifier("""This is great !""" , top_k=2 ) self.assertEqual( nested_simplify(A ) , [{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}] ) _SCREAMING_SNAKE_CASE = text_classifier(["""This is great !""", """This is bad"""] , top_k=2 ) self.assertEqual( nested_simplify(A ) , [ [{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}], [{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}], ] , ) _SCREAMING_SNAKE_CASE = text_classifier("""This is great !""" , top_k=1 ) self.assertEqual(nested_simplify(A ) , [{"""label""": """LABEL_0""", """score""": 0.504}] ) # Legacy behavior _SCREAMING_SNAKE_CASE = text_classifier("""This is great !""" , return_all_scores=A ) self.assertEqual(nested_simplify(A ) , [{"""label""": """LABEL_0""", """score""": 0.504}] ) _SCREAMING_SNAKE_CASE = text_classifier("""This is great !""" , return_all_scores=A ) self.assertEqual( nested_simplify(A ) , [[{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}]] ) _SCREAMING_SNAKE_CASE = text_classifier(["""This is great !""", """Something else"""] , return_all_scores=A ) self.assertEqual( nested_simplify(A ) , [ [{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}], [{"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_1""", """score""": 0.496}], ] , ) _SCREAMING_SNAKE_CASE = text_classifier(["""This is great !""", """Something else"""] , return_all_scores=A ) self.assertEqual( nested_simplify(A ) , [ {"""label""": """LABEL_0""", """score""": 0.504}, {"""label""": """LABEL_0""", """score""": 0.504}, ] , ) @require_torch def snake_case_( self ) -> int: import torch _SCREAMING_SNAKE_CASE = pipeline( task="""text-classification""" , model="""hf-internal-testing/tiny-random-distilbert""" , framework="""pt""" , device=torch.device("""cpu""" ) , ) _SCREAMING_SNAKE_CASE = text_classifier("""This is great !""" ) self.assertEqual(nested_simplify(A ) , [{"""label""": """LABEL_0""", """score""": 0.504}] ) @require_tf def snake_case_( self ) -> List[Any]: _SCREAMING_SNAKE_CASE = pipeline( task="""text-classification""" , model="""hf-internal-testing/tiny-random-distilbert""" , framework="""tf""" ) _SCREAMING_SNAKE_CASE = text_classifier("""This is great !""" ) self.assertEqual(nested_simplify(A ) , [{"""label""": """LABEL_0""", """score""": 0.504}] ) @slow @require_torch def snake_case_( self ) -> Any: _SCREAMING_SNAKE_CASE = pipeline("""text-classification""" ) _SCREAMING_SNAKE_CASE = text_classifier("""This is great !""" ) self.assertEqual(nested_simplify(A ) , [{"""label""": """POSITIVE""", """score""": 1.0}] ) _SCREAMING_SNAKE_CASE = text_classifier("""This is bad !""" ) self.assertEqual(nested_simplify(A ) , [{"""label""": """NEGATIVE""", """score""": 1.0}] ) _SCREAMING_SNAKE_CASE = text_classifier("""Birds are a type of animal""" ) self.assertEqual(nested_simplify(A ) , [{"""label""": """POSITIVE""", """score""": 0.988}] ) @slow @require_tf def snake_case_( self ) -> List[Any]: _SCREAMING_SNAKE_CASE = pipeline("""text-classification""" , framework="""tf""" ) _SCREAMING_SNAKE_CASE = text_classifier("""This is great !""" ) self.assertEqual(nested_simplify(A ) , [{"""label""": """POSITIVE""", """score""": 1.0}] ) _SCREAMING_SNAKE_CASE = text_classifier("""This is bad !""" ) self.assertEqual(nested_simplify(A ) , [{"""label""": """NEGATIVE""", """score""": 1.0}] ) _SCREAMING_SNAKE_CASE = text_classifier("""Birds are a type of animal""" ) self.assertEqual(nested_simplify(A ) , [{"""label""": """POSITIVE""", """score""": 0.988}] ) def snake_case_( self , A , A , A ) -> Optional[Any]: _SCREAMING_SNAKE_CASE = TextClassificationPipeline(model=A , tokenizer=A ) return text_classifier, ["HuggingFace is in", "This is another test"] def snake_case_( self , A , A ) -> int: _SCREAMING_SNAKE_CASE = text_classifier.model # Small inputs because BartTokenizer tiny has maximum position embeddings = 22 _SCREAMING_SNAKE_CASE = """HuggingFace is in""" _SCREAMING_SNAKE_CASE = text_classifier(A ) self.assertEqual(nested_simplify(A ) , [{"""label""": ANY(A ), """score""": ANY(A )}] ) self.assertTrue(outputs[0]["""label"""] in model.config.idalabel.values() ) _SCREAMING_SNAKE_CASE = ["""HuggingFace is in """, """Paris is in France"""] _SCREAMING_SNAKE_CASE = text_classifier(A ) self.assertEqual( nested_simplify(A ) , [{"""label""": ANY(A ), """score""": ANY(A )}, {"""label""": ANY(A ), """score""": ANY(A )}] , ) self.assertTrue(outputs[0]["""label"""] in model.config.idalabel.values() ) self.assertTrue(outputs[1]["""label"""] in model.config.idalabel.values() ) # Forcing to get all results with `top_k=None` # This is NOT the legacy format _SCREAMING_SNAKE_CASE = text_classifier(A , top_k=A ) _SCREAMING_SNAKE_CASE = len(model.config.idalabel.values() ) self.assertEqual( nested_simplify(A ) , [[{"""label""": ANY(A ), """score""": ANY(A )}] * N, [{"""label""": ANY(A ), """score""": ANY(A )}] * N] , ) _SCREAMING_SNAKE_CASE = {"""text""": """HuggingFace is in """, """text_pair""": """Paris is in France"""} _SCREAMING_SNAKE_CASE = text_classifier(A ) self.assertEqual( nested_simplify(A ) , {"""label""": ANY(A ), """score""": ANY(A )} , ) self.assertTrue(outputs["""label"""] in model.config.idalabel.values() ) # This might be used a text pair, but tokenizer + pipe interaction # makes it hard to understand that it's not using the pair properly # https://github.com/huggingface/transformers/issues/17305 # We disabled this usage instead as it was outputting wrong outputs. _SCREAMING_SNAKE_CASE = [["""HuggingFace is in """, """Paris is in France"""]] with self.assertRaises(A ): text_classifier(A ) # This used to be valid for doing text pairs # We're keeping it working because of backward compatibility _SCREAMING_SNAKE_CASE = text_classifier([[["""HuggingFace is in """, """Paris is in France"""]]] ) self.assertEqual( nested_simplify(A ) , [{"""label""": ANY(A ), """score""": ANY(A )}] , ) self.assertTrue(outputs[0]["""label"""] in model.config.idalabel.values() )
58
def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : int ): return [sentence[i : i + ngram_size] for i in range(len(lowerCAmelCase_ ) - ngram_size + 1 )] if __name__ == "__main__": from doctest import testmod testmod()
284
0
def __lowerCamelCase ( lowerCamelCase__ : int = 100 ): '''simple docstring''' lowerCamelCase = (n * (n + 1) // 2) ** 2 lowerCamelCase = n * (n + 1) * (2 * n + 1) // 6 return sum_cubes - sum_squares if __name__ == "__main__": print(f"""{solution() = }""")
365
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase : Union[str, Any] = { "configuration_mctct": ["MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP", "MCTCTConfig"], "feature_extraction_mctct": ["MCTCTFeatureExtractor"], "processing_mctct": ["MCTCTProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : List[Any] = [ "MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST", "MCTCTForCTC", "MCTCTModel", "MCTCTPreTrainedModel", ] if TYPE_CHECKING: from .configuration_mctct import MCTCT_PRETRAINED_CONFIG_ARCHIVE_MAP, MCTCTConfig from .feature_extraction_mctct import MCTCTFeatureExtractor from .processing_mctct import MCTCTProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mctct import MCTCT_PRETRAINED_MODEL_ARCHIVE_LIST, MCTCTForCTC, MCTCTModel, MCTCTPreTrainedModel else: import sys UpperCAmelCase : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
0
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from argparse import ArgumentParser from accelerate.commands.config import get_config_parser from accelerate.commands.env import env_command_parser from accelerate.commands.launch import launch_command_parser from accelerate.commands.test import test_command_parser from accelerate.commands.tpu import tpu_command_parser def __SCREAMING_SNAKE_CASE ( ): '''simple docstring''' _UpperCAmelCase = ArgumentParser("Accelerate CLI tool" , usage="accelerate <command> [<args>]" , allow_abbrev=snake_case_ ) _UpperCAmelCase = parser.add_subparsers(help="accelerate command helpers" ) # Register commands get_config_parser(subparsers=snake_case_ ) env_command_parser(subparsers=snake_case_ ) launch_command_parser(subparsers=snake_case_ ) tpu_command_parser(subparsers=snake_case_ ) test_command_parser(subparsers=snake_case_ ) # Let's go _UpperCAmelCase = parser.parse_args() if not hasattr(snake_case_ , "func" ): parser.print_help() exit(1 ) # Run args.func(snake_case_ ) if __name__ == "__main__": main()
133
def __SCREAMING_SNAKE_CASE ( snake_case_ = 1000 ): '''simple docstring''' _UpperCAmelCase = 2**power _UpperCAmelCase = 0 while n: _UpperCAmelCase , _UpperCAmelCase = r + n % 10, n // 10 return r if __name__ == "__main__": print(solution(int(str(input()).strip())))
133
1
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( BertTokenizer, ViltConfig, ViltForImageAndTextRetrieval, ViltForImagesAndTextClassification, ViltForMaskedLM, ViltForQuestionAnswering, ViltImageProcessor, ViltProcessor, ) from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase_ : Dict = logging.get_logger(__name__) def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase=False , lowerCAmelCase=False , lowerCAmelCase=False ): '''simple docstring''' UpperCAmelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'''transformer.blocks.{i}.norm1.weight''', F'''vilt.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((F'''transformer.blocks.{i}.norm1.bias''', F'''vilt.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append( (F'''transformer.blocks.{i}.attn.proj.weight''', F'''vilt.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append( (F'''transformer.blocks.{i}.attn.proj.bias''', F'''vilt.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((F'''transformer.blocks.{i}.norm2.weight''', F'''vilt.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((F'''transformer.blocks.{i}.norm2.bias''', F'''vilt.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append( (F'''transformer.blocks.{i}.mlp.fc1.weight''', F'''vilt.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((F'''transformer.blocks.{i}.mlp.fc1.bias''', F'''vilt.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((F'''transformer.blocks.{i}.mlp.fc2.weight''', F'''vilt.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((F'''transformer.blocks.{i}.mlp.fc2.bias''', F'''vilt.encoder.layer.{i}.output.dense.bias''') ) # embeddings rename_keys.extend( [ # text embeddings ("""text_embeddings.word_embeddings.weight""", """vilt.embeddings.text_embeddings.word_embeddings.weight"""), ( """text_embeddings.position_embeddings.weight""", """vilt.embeddings.text_embeddings.position_embeddings.weight""", ), ("""text_embeddings.position_ids""", """vilt.embeddings.text_embeddings.position_ids"""), ( """text_embeddings.token_type_embeddings.weight""", """vilt.embeddings.text_embeddings.token_type_embeddings.weight""", ), ("""text_embeddings.LayerNorm.weight""", """vilt.embeddings.text_embeddings.LayerNorm.weight"""), ("""text_embeddings.LayerNorm.bias""", """vilt.embeddings.text_embeddings.LayerNorm.bias"""), # patch embeddings ("""transformer.cls_token""", """vilt.embeddings.cls_token"""), ("""transformer.patch_embed.proj.weight""", """vilt.embeddings.patch_embeddings.projection.weight"""), ("""transformer.patch_embed.proj.bias""", """vilt.embeddings.patch_embeddings.projection.bias"""), ("""transformer.pos_embed""", """vilt.embeddings.position_embeddings"""), # token type embeddings ("""token_type_embeddings.weight""", """vilt.embeddings.token_type_embeddings.weight"""), ] ) # final layernorm + pooler rename_keys.extend( [ ("""transformer.norm.weight""", """vilt.layernorm.weight"""), ("""transformer.norm.bias""", """vilt.layernorm.bias"""), ("""pooler.dense.weight""", """vilt.pooler.dense.weight"""), ("""pooler.dense.bias""", """vilt.pooler.dense.bias"""), ] ) # classifier head(s) if vqa_model: # classification head rename_keys.extend( [ ("""vqa_classifier.0.weight""", """classifier.0.weight"""), ("""vqa_classifier.0.bias""", """classifier.0.bias"""), ("""vqa_classifier.1.weight""", """classifier.1.weight"""), ("""vqa_classifier.1.bias""", """classifier.1.bias"""), ("""vqa_classifier.3.weight""", """classifier.3.weight"""), ("""vqa_classifier.3.bias""", """classifier.3.bias"""), ] ) elif nlvr_model: # classification head rename_keys.extend( [ ("""nlvr2_classifier.0.weight""", """classifier.0.weight"""), ("""nlvr2_classifier.0.bias""", """classifier.0.bias"""), ("""nlvr2_classifier.1.weight""", """classifier.1.weight"""), ("""nlvr2_classifier.1.bias""", """classifier.1.bias"""), ("""nlvr2_classifier.3.weight""", """classifier.3.weight"""), ("""nlvr2_classifier.3.bias""", """classifier.3.bias"""), ] ) else: pass return rename_keys def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' for i in range(config.num_hidden_layers ): UpperCAmelCase = """vilt.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase = state_dict.pop(F'''transformer.blocks.{i}.attn.qkv.weight''' ) UpperCAmelCase = state_dict.pop(F'''transformer.blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase = in_proj_weight[ : config.hidden_size, : ] UpperCAmelCase = in_proj_bias[: config.hidden_size] UpperCAmelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase = in_proj_bias[-config.hidden_size :] def _lowerCAmelCase ( lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = ["""head.weight""", """head.bias"""] for k in ignore_keys: state_dict.pop(lowerCAmelCase , lowerCAmelCase ) def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = dct.pop(lowerCAmelCase ) UpperCAmelCase = val @torch.no_grad() def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' UpperCAmelCase = ViltConfig(image_size=384 , patch_size=32 , tie_word_embeddings=lowerCAmelCase ) UpperCAmelCase = False UpperCAmelCase = False UpperCAmelCase = False UpperCAmelCase = False if "vqa" in checkpoint_url: UpperCAmelCase = True UpperCAmelCase = 3129 UpperCAmelCase = """huggingface/label-files""" UpperCAmelCase = """vqa2-id2label.json""" UpperCAmelCase = json.load(open(hf_hub_download(lowerCAmelCase , lowerCAmelCase , repo_type="""dataset""" ) , """r""" ) ) UpperCAmelCase = {int(lowerCAmelCase ): v for k, v in idalabel.items()} UpperCAmelCase = idalabel UpperCAmelCase = {v: k for k, v in idalabel.items()} UpperCAmelCase = ViltForQuestionAnswering(lowerCAmelCase ) elif "nlvr" in checkpoint_url: UpperCAmelCase = True UpperCAmelCase = 2 UpperCAmelCase = {0: """False""", 1: """True"""} UpperCAmelCase = {v: k for k, v in config.idalabel.items()} UpperCAmelCase = 3 UpperCAmelCase = ViltForImagesAndTextClassification(lowerCAmelCase ) elif "irtr" in checkpoint_url: UpperCAmelCase = True UpperCAmelCase = ViltForImageAndTextRetrieval(lowerCAmelCase ) elif "mlm_itm" in checkpoint_url: UpperCAmelCase = True UpperCAmelCase = ViltForMaskedLM(lowerCAmelCase ) else: raise ValueError("""Unknown model type""" ) # load state_dict of original model, remove and rename some keys UpperCAmelCase = torch.hub.load_state_dict_from_url(lowerCAmelCase , map_location="""cpu""" )["""state_dict"""] UpperCAmelCase = create_rename_keys(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) for src, dest in rename_keys: rename_key(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) read_in_q_k_v(lowerCAmelCase , lowerCAmelCase ) if mlm_model or irtr_model: UpperCAmelCase = ["""itm_score.fc.weight""", """itm_score.fc.bias"""] for k in ignore_keys: state_dict.pop(lowerCAmelCase , lowerCAmelCase ) # load state dict into HuggingFace model model.eval() if mlm_model: UpperCAmelCase , UpperCAmelCase = model.load_state_dict(lowerCAmelCase , strict=lowerCAmelCase ) assert missing_keys == ["mlm_score.decoder.bias"] else: model.load_state_dict(lowerCAmelCase ) # Define processor UpperCAmelCase = ViltImageProcessor(size=384 ) UpperCAmelCase = BertTokenizer.from_pretrained("""bert-base-uncased""" ) UpperCAmelCase = ViltProcessor(lowerCAmelCase , lowerCAmelCase ) # Forward pass on example inputs (image + text) if nlvr_model: UpperCAmelCase = Image.open(requests.get("""https://lil.nlp.cornell.edu/nlvr/exs/ex0_0.jpg""" , stream=lowerCAmelCase ).raw ) UpperCAmelCase = Image.open(requests.get("""https://lil.nlp.cornell.edu/nlvr/exs/ex0_0.jpg""" , stream=lowerCAmelCase ).raw ) UpperCAmelCase = ( """The left image contains twice the number of dogs as the right image, and at least two dogs in total are""" """ standing.""" ) UpperCAmelCase = processor(lowerCAmelCase , lowerCAmelCase , return_tensors="""pt""" ) UpperCAmelCase = processor(lowerCAmelCase , lowerCAmelCase , return_tensors="""pt""" ) UpperCAmelCase = model( input_ids=encoding_a.input_ids , pixel_values=encoding_a.pixel_values , pixel_values_a=encoding_a.pixel_values , ) else: UpperCAmelCase = Image.open(requests.get("""http://images.cocodataset.org/val2017/000000039769.jpg""" , stream=lowerCAmelCase ).raw ) if mlm_model: UpperCAmelCase = """a bunch of [MASK] laying on a [MASK].""" else: UpperCAmelCase = """How many cats are there?""" UpperCAmelCase = processor(lowerCAmelCase , lowerCAmelCase , return_tensors="""pt""" ) UpperCAmelCase = model(**lowerCAmelCase ) # Verify outputs if mlm_model: UpperCAmelCase = torch.Size([1, 11, 30522] ) UpperCAmelCase = torch.tensor([-12.50_61, -12.51_23, -12.51_74] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, 0, :3] , lowerCAmelCase , atol=1e-4 ) # verify masked token prediction equals "cats" UpperCAmelCase = outputs.logits[0, 4, :].argmax(-1 ).item() assert tokenizer.decode([predicted_id] ) == "cats" elif vqa_model: UpperCAmelCase = torch.Size([1, 3129] ) UpperCAmelCase = torch.tensor([-15.94_95, -18.14_72, -10.30_41] ) assert torch.allclose(outputs.logits[0, :3] , lowerCAmelCase , atol=1e-4 ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, 0, :3] , lowerCAmelCase , atol=1e-4 ) # verify vqa prediction equals "2" UpperCAmelCase = outputs.logits.argmax(-1 ).item() assert model.config.idalabel[predicted_idx] == "2" elif nlvr_model: UpperCAmelCase = torch.Size([1, 2] ) UpperCAmelCase = torch.tensor([-2.87_21, 2.12_91] ) assert torch.allclose(outputs.logits[0, :3] , lowerCAmelCase , atol=1e-4 ) assert outputs.logits.shape == expected_shape Path(lowerCAmelCase ).mkdir(exist_ok=lowerCAmelCase ) print(F'''Saving model and processor to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowerCAmelCase ) processor.save_pretrained(lowerCAmelCase ) if __name__ == "__main__": lowerCAmelCase_ : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://github.com/dandelin/ViLT/releases/download/200k/vilt_200k_mlm_itm.ckpt''', type=str, help='''URL of the checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) lowerCAmelCase_ : Optional[Any] = parser.parse_args() convert_vilt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
248
"""simple docstring""" import math import random from typing import Any from .hill_climbing import SearchProblem def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase = True , lowerCAmelCase = math.inf , lowerCAmelCase = -math.inf , lowerCAmelCase = math.inf , lowerCAmelCase = -math.inf , lowerCAmelCase = False , lowerCAmelCase = 100 , lowerCAmelCase = 0.01 , lowerCAmelCase = 1 , ): '''simple docstring''' UpperCAmelCase = False UpperCAmelCase = search_prob UpperCAmelCase = start_temperate UpperCAmelCase = [] UpperCAmelCase = 0 UpperCAmelCase = None while not search_end: UpperCAmelCase = current_state.score() if best_state is None or current_score > best_state.score(): UpperCAmelCase = current_state scores.append(lowerCAmelCase ) iterations += 1 UpperCAmelCase = None UpperCAmelCase = current_state.get_neighbors() while ( next_state is None and neighbors ): # till we do not find a neighbor that we can move to UpperCAmelCase = random.randint(0 , len(lowerCAmelCase ) - 1 ) # picking a random neighbor UpperCAmelCase = neighbors.pop(lowerCAmelCase ) UpperCAmelCase = picked_neighbor.score() - current_score if ( picked_neighbor.x > max_x or picked_neighbor.x < min_x or picked_neighbor.y > max_y or picked_neighbor.y < min_y ): continue # neighbor outside our bounds if not find_max: UpperCAmelCase = change * -1 # in case we are finding minimum if change > 0: # improves the solution UpperCAmelCase = picked_neighbor else: UpperCAmelCase = (math.e) ** ( change / current_temp ) # probability generation function if random.random() < probability: # random number within probability UpperCAmelCase = picked_neighbor UpperCAmelCase = current_temp - (current_temp * rate_of_decrease) if current_temp < threshold_temp or next_state is None: # temperature below threshold, or could not find a suitable neighbor UpperCAmelCase = True else: UpperCAmelCase = next_state if visualization: from matplotlib import pyplot as plt plt.plot(range(lowerCAmelCase ) , lowerCAmelCase ) plt.xlabel("""Iterations""" ) plt.ylabel("""Function values""" ) plt.show() return best_state if __name__ == "__main__": def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' return (x**2) + (y**2) # starting the problem with initial coordinates (12, 47) lowerCAmelCase_ : List[str] = SearchProblem(x=1_2, y=4_7, step_size=1, function_to_optimize=test_fa) lowerCAmelCase_ : List[str] = simulated_annealing( prob, find_max=False, max_x=1_0_0, min_x=5, max_y=5_0, min_y=-5, visualization=True ) print( '''The minimum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 ''' F'and 50 > y > - 5 found via hill climbing: {local_min.score()}' ) # starting the problem with initial coordinates (12, 47) lowerCAmelCase_ : int = SearchProblem(x=1_2, y=4_7, step_size=1, function_to_optimize=test_fa) lowerCAmelCase_ : Optional[Any] = simulated_annealing( prob, find_max=True, max_x=1_0_0, min_x=5, max_y=5_0, min_y=-5, visualization=True ) print( '''The maximum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 ''' F'and 50 > y > - 5 found via hill climbing: {local_min.score()}' ) def _lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase ): '''simple docstring''' return (3 * x**2) - (6 * y) lowerCAmelCase_ : Dict = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) lowerCAmelCase_ : List[Any] = simulated_annealing(prob, find_max=False, visualization=True) print( '''The minimum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: ''' F'{local_min.score()}' ) lowerCAmelCase_ : List[str] = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) lowerCAmelCase_ : List[Any] = simulated_annealing(prob, find_max=True, visualization=True) print( '''The maximum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: ''' F'{local_min.score()}' )
248
1
"""simple docstring""" from pathlib import Path import numpy as np from PIL import Image def _lowerCAmelCase ( lowercase_ ): UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = rgb[:, :, 0], rgb[:, :, 1], rgb[:, :, 2] return 0.2_9_8_9 * r + 0.5_8_7_0 * g + 0.1_1_4_0 * b def _lowerCAmelCase ( lowercase_ ): return (gray > 127) & (gray <= 255) def _lowerCAmelCase ( lowercase_ , lowercase_ ): UpperCAmelCase = np.zeros_like(lowercase_ ) UpperCAmelCase = np.zeros( (image.shape[0] + kernel.shape[0] - 1, image.shape[1] + kernel.shape[1] - 1) ) # Copy image to padded image UpperCAmelCase = image # Iterate over image & apply kernel for x in range(image.shape[1] ): for y in range(image.shape[0] ): UpperCAmelCase = ( kernel * image_padded[y : y + kernel.shape[0], x : x + kernel.shape[1]] ).sum() UpperCAmelCase = int(summation > 0 ) return output if __name__ == "__main__": # read original image snake_case_ = Path(__file__).resolve().parent / """image_data""" / """lena.jpg""" snake_case_ = np.array(Image.open(lena_path)) # kernel to be applied snake_case_ = np.array([[0, 1, 0], [1, 1, 1], [0, 1, 0]]) snake_case_ = dilation(gray_to_binary(rgb_to_gray(lena)), structuring_element) # Save the output image snake_case_ = Image.fromarray(output).convert("""RGB""") pil_img.save("""result_dilation.png""")
78
import os from distutils.util import strtobool def __lowercase ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Tuple: '''simple docstring''' for e in env_keys: SCREAMING_SNAKE_CASE = int(os.environ.get(_SCREAMING_SNAKE_CASE , -1 ) ) if val >= 0: return val return default def __lowercase ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False ) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE = os.environ.get(_SCREAMING_SNAKE_CASE , str(_SCREAMING_SNAKE_CASE ) ) return strtobool(_SCREAMING_SNAKE_CASE ) == 1 # As its name indicates `strtobool` actually returns an int... def __lowercase ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE="no" ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE = os.environ.get(_SCREAMING_SNAKE_CASE , str(_SCREAMING_SNAKE_CASE ) ) return value
296
0
# This model implementation is heavily inspired by https://github.com/haofanwang/ControlNet-for-Diffusers/ import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, ControlNetModel, DDIMScheduler, StableDiffusionControlNetImgaImgPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet import MultiControlNetModel from diffusers.utils import floats_tensor, load_image, load_numpy, randn_tensor, slow, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, ) enable_full_determinism() class lowercase__ ( a_ , a_ , a_ , unittest.TestCase ): A__ : Any =StableDiffusionControlNetImgaImgPipeline A__ : int =TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''height''', '''width'''} A__ : List[str] =TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS A__ : int =IMAGE_TO_IMAGE_IMAGE_PARAMS.union({"""control_image"""} ) A__ : Union[str, Any] =IMAGE_TO_IMAGE_IMAGE_PARAMS def A_ ( self : List[Any] ): torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=lowercase_ , set_alpha_to_one=lowercase_ , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) SCREAMING_SNAKE_CASE__ = CLIPTextModel(lowercase_ ) SCREAMING_SNAKE_CASE__ = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) SCREAMING_SNAKE_CASE__ = { 'unet': unet, 'controlnet': controlnet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def A_ ( self : Optional[int] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Dict=0 ): if str(lowercase_ ).startswith('mps' ): SCREAMING_SNAKE_CASE__ = torch.manual_seed(lowercase_ ) else: SCREAMING_SNAKE_CASE__ = torch.Generator(device=lowercase_ ).manual_seed(lowercase_ ) SCREAMING_SNAKE_CASE__ = 2 SCREAMING_SNAKE_CASE__ = randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=lowercase_ , device=torch.device(lowercase_ ) , ) SCREAMING_SNAKE_CASE__ = floats_tensor(control_image.shape , rng=random.Random(lowercase_ ) ).to(lowercase_ ) SCREAMING_SNAKE_CASE__ = image.cpu().permute(0 , 2 , 3 , 1 )[0] SCREAMING_SNAKE_CASE__ = Image.fromarray(np.uinta(lowercase_ ) ).convert('RGB' ).resize((64, 64) ) SCREAMING_SNAKE_CASE__ = { 'prompt': 'A painting of a squirrel eating a burger', 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'output_type': 'numpy', 'image': image, 'control_image': control_image, } return inputs def A_ ( self : Any ): return self._test_attention_slicing_forward_pass(expected_max_diff=2e-3 ) @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def A_ ( self : Any ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2e-3 ) def A_ ( self : Dict ): self._test_inference_batch_single_identical(expected_max_diff=2e-3 ) class lowercase__ ( a_ , a_ , unittest.TestCase ): A__ : Union[str, Any] =StableDiffusionControlNetImgaImgPipeline A__ : Union[str, Any] =TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''height''', '''width'''} A__ : Tuple =TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS A__ : Optional[int] =frozenset([] ) # TO_DO: add image_params once refactored VaeImageProcessor.preprocess def A_ ( self : Tuple ): torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , ) torch.manual_seed(0 ) def init_weights(UpperCAmelCase_ : Optional[int] ): if isinstance(lowercase_ , torch.nn.Convad ): torch.nn.init.normal(m.weight ) m.bias.data.fill_(1.0 ) SCREAMING_SNAKE_CASE__ = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(lowercase_ ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(lowercase_ ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=lowercase_ , set_alpha_to_one=lowercase_ , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) SCREAMING_SNAKE_CASE__ = CLIPTextModel(lowercase_ ) SCREAMING_SNAKE_CASE__ = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) SCREAMING_SNAKE_CASE__ = MultiControlNetModel([controlneta, controlneta] ) SCREAMING_SNAKE_CASE__ = { 'unet': unet, 'controlnet': controlnet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def A_ ( self : Dict , UpperCAmelCase_ : int , UpperCAmelCase_ : Dict=0 ): if str(lowercase_ ).startswith('mps' ): SCREAMING_SNAKE_CASE__ = torch.manual_seed(lowercase_ ) else: SCREAMING_SNAKE_CASE__ = torch.Generator(device=lowercase_ ).manual_seed(lowercase_ ) SCREAMING_SNAKE_CASE__ = 2 SCREAMING_SNAKE_CASE__ = [ randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=lowercase_ , device=torch.device(lowercase_ ) , ), randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=lowercase_ , device=torch.device(lowercase_ ) , ), ] SCREAMING_SNAKE_CASE__ = floats_tensor(control_image[0].shape , rng=random.Random(lowercase_ ) ).to(lowercase_ ) SCREAMING_SNAKE_CASE__ = image.cpu().permute(0 , 2 , 3 , 1 )[0] SCREAMING_SNAKE_CASE__ = Image.fromarray(np.uinta(lowercase_ ) ).convert('RGB' ).resize((64, 64) ) SCREAMING_SNAKE_CASE__ = { 'prompt': 'A painting of a squirrel eating a burger', 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'output_type': 'numpy', 'image': image, 'control_image': control_image, } return inputs def A_ ( self : Optional[int] ): SCREAMING_SNAKE_CASE__ = self.get_dummy_components() SCREAMING_SNAKE_CASE__ = self.pipeline_class(**lowercase_ ) pipe.to(lowercase_ ) SCREAMING_SNAKE_CASE__ = 10.0 SCREAMING_SNAKE_CASE__ = 4 SCREAMING_SNAKE_CASE__ = self.get_dummy_inputs(lowercase_ ) SCREAMING_SNAKE_CASE__ = steps SCREAMING_SNAKE_CASE__ = scale SCREAMING_SNAKE_CASE__ = pipe(**lowercase_ )[0] SCREAMING_SNAKE_CASE__ = self.get_dummy_inputs(lowercase_ ) SCREAMING_SNAKE_CASE__ = steps SCREAMING_SNAKE_CASE__ = scale SCREAMING_SNAKE_CASE__ = pipe(**lowercase_ , control_guidance_start=0.1 , control_guidance_end=0.2 )[0] SCREAMING_SNAKE_CASE__ = self.get_dummy_inputs(lowercase_ ) SCREAMING_SNAKE_CASE__ = steps SCREAMING_SNAKE_CASE__ = scale SCREAMING_SNAKE_CASE__ = pipe(**lowercase_ , control_guidance_start=[0.1, 0.3] , control_guidance_end=[0.2, 0.7] )[0] SCREAMING_SNAKE_CASE__ = self.get_dummy_inputs(lowercase_ ) SCREAMING_SNAKE_CASE__ = steps SCREAMING_SNAKE_CASE__ = scale SCREAMING_SNAKE_CASE__ = pipe(**lowercase_ , control_guidance_start=0.4 , control_guidance_end=[0.5, 0.8] )[0] # make sure that all outputs are different assert np.sum(np.abs(output_a - output_a ) ) > 1e-3 assert np.sum(np.abs(output_a - output_a ) ) > 1e-3 assert np.sum(np.abs(output_a - output_a ) ) > 1e-3 def A_ ( self : Union[str, Any] ): return self._test_attention_slicing_forward_pass(expected_max_diff=2e-3 ) @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def A_ ( self : Any ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2e-3 ) def A_ ( self : Tuple ): self._test_inference_batch_single_identical(expected_max_diff=2e-3 ) def A_ ( self : List[Any] ): SCREAMING_SNAKE_CASE__ = self.get_dummy_components() SCREAMING_SNAKE_CASE__ = self.pipeline_class(**lowercase_ ) pipe.to(lowercase_ ) pipe.set_progress_bar_config(disable=lowercase_ ) with tempfile.TemporaryDirectory() as tmpdir: try: # save_pretrained is not implemented for Multi-ControlNet pipe.save_pretrained(lowercase_ ) except NotImplementedError: pass @slow @require_torch_gpu class lowercase__ ( unittest.TestCase ): def A_ ( self : Optional[int] ): super().tearDown() gc.collect() torch.cuda.empty_cache() def A_ ( self : int ): SCREAMING_SNAKE_CASE__ = ControlNetModel.from_pretrained('lllyasviel/sd-controlnet-canny' ) SCREAMING_SNAKE_CASE__ = StableDiffusionControlNetImgaImgPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' , safety_checker=lowercase_ , controlnet=lowercase_ ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=lowercase_ ) SCREAMING_SNAKE_CASE__ = torch.Generator(device='cpu' ).manual_seed(0 ) SCREAMING_SNAKE_CASE__ = 'evil space-punk bird' SCREAMING_SNAKE_CASE__ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png' ).resize((512, 512) ) SCREAMING_SNAKE_CASE__ = load_image( 'https://huggingface.co/lllyasviel/sd-controlnet-canny/resolve/main/images/bird.png' ).resize((512, 512) ) SCREAMING_SNAKE_CASE__ = pipe( lowercase_ , lowercase_ , control_image=lowercase_ , generator=lowercase_ , output_type='np' , num_inference_steps=50 , strength=0.6 , ) SCREAMING_SNAKE_CASE__ = output.images[0] assert image.shape == (512, 512, 3) SCREAMING_SNAKE_CASE__ = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/img2img.npy' ) assert np.abs(expected_image - image ).max() < 9e-2
364
import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_=None ) -> Optional[Any]: '''simple docstring''' assert torch_layer.weight.shape == weight.shape, F'{torch_layer} layer.weight does not match' SCREAMING_SNAKE_CASE__ = nn.Parameter(UpperCamelCase_ ) if bias is not None: assert torch_layer.bias.shape == bias.shape, F'{torch_layer} layer.bias does not match' SCREAMING_SNAKE_CASE__ = nn.Parameter(UpperCamelCase_ ) def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE__ = np.asarray(weights[0] ) SCREAMING_SNAKE_CASE__ = np.asarray(weights[1] ) SCREAMING_SNAKE_CASE__ = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key , torch.tensor(UpperCamelCase_ ).transpose(1 , 2 ).contiguous().view(-1 , UpperCamelCase_ ) , ) set_param( torch_layer.self_attention.value , torch.tensor(UpperCamelCase_ ).transpose(1 , 2 ).contiguous().view(-1 , UpperCamelCase_ ) , ) set_param( torch_layer.output.dense , torch.tensor(UpperCamelCase_ ).view(-1 , UpperCamelCase_ ).contiguous().transpose(0 , 1 ) , ) def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE__ = np.asarray(weights[0] ) SCREAMING_SNAKE_CASE__ = np.asarray(weights[1] ) SCREAMING_SNAKE_CASE__ = np.asarray(weights[2] ) SCREAMING_SNAKE_CASE__ = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query , torch.tensor(UpperCamelCase_ ).transpose(1 , 2 ).contiguous().view(-1 , UpperCamelCase_ ) , ) set_param( torch_layer.self_attention.key , torch.tensor(UpperCamelCase_ ).transpose(1 , 2 ).contiguous().view(-1 , UpperCamelCase_ ) , ) set_param( torch_layer.self_attention.value , torch.tensor(UpperCamelCase_ ).transpose(1 , 2 ).contiguous().view(-1 , UpperCamelCase_ ) , ) set_param( torch_layer.output.dense , torch.tensor(UpperCamelCase_ ).view(-1 , UpperCamelCase_ ).contiguous().transpose(0 , 1 ) , ) def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> Tuple: '''simple docstring''' SCREAMING_SNAKE_CASE__ = weights[0][0][0] SCREAMING_SNAKE_CASE__ = np.asarray(layer_norm_a[0] ) SCREAMING_SNAKE_CASE__ = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm , torch.tensor(UpperCamelCase_ ) , torch.tensor(UpperCamelCase_ ) , ) # lsh weights + output SCREAMING_SNAKE_CASE__ = weights[0][1] if len(UpperCamelCase_ ) < 4: set_layer_weights_in_torch_lsh(UpperCamelCase_ , torch_block.attention , UpperCamelCase_ ) else: set_layer_weights_in_torch_local(UpperCamelCase_ , torch_block.attention , UpperCamelCase_ ) # intermediate weighs SCREAMING_SNAKE_CASE__ = weights[2][0][1][2] # Chunked Feed Forward if len(UpperCamelCase_ ) == 4: SCREAMING_SNAKE_CASE__ = intermediate_weights[2] # layernorm 2 SCREAMING_SNAKE_CASE__ = np.asarray(intermediate_weights[0][0] ) SCREAMING_SNAKE_CASE__ = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm , torch.tensor(UpperCamelCase_ ) , torch.tensor(UpperCamelCase_ ) , ) # intermediate dense SCREAMING_SNAKE_CASE__ = np.asarray(intermediate_weights[1][0] ) SCREAMING_SNAKE_CASE__ = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense , torch.tensor(UpperCamelCase_ ).transpose(0 , 1 ).contiguous() , torch.tensor(UpperCamelCase_ ) , ) # intermediate out SCREAMING_SNAKE_CASE__ = np.asarray(intermediate_weights[4][0] ) SCREAMING_SNAKE_CASE__ = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense , torch.tensor(UpperCamelCase_ ).transpose(0 , 1 ).contiguous() , torch.tensor(UpperCamelCase_ ) , ) def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE__ = torch_model.reformer # word embeds SCREAMING_SNAKE_CASE__ = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings , torch.tensor(UpperCamelCase_ ) , ) if isinstance(weights[3] , UpperCamelCase_ ): SCREAMING_SNAKE_CASE__ = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): SCREAMING_SNAKE_CASE__ = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), F'{position_embeddings[emb_idx]} emb does not match' SCREAMING_SNAKE_CASE__ = nn.Parameter(torch.tensor(UpperCamelCase_ ) ) SCREAMING_SNAKE_CASE__ = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( UpperCamelCase_ ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): SCREAMING_SNAKE_CASE__ = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) # output layer norm SCREAMING_SNAKE_CASE__ = np.asarray(weights[7][0] ) SCREAMING_SNAKE_CASE__ = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm , torch.tensor(UpperCamelCase_ ) , torch.tensor(UpperCamelCase_ ) , ) # output embeddings SCREAMING_SNAKE_CASE__ = np.asarray(weights[9][0] ) SCREAMING_SNAKE_CASE__ = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder , torch.tensor(UpperCamelCase_ ).transpose(0 , 1 ).contiguous() , torch.tensor(UpperCamelCase_ ) , ) def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE__ = ReformerConfig.from_json_file(UpperCamelCase_ ) print(F'Building PyTorch model from configuration: {config}' ) SCREAMING_SNAKE_CASE__ = ReformerModelWithLMHead(UpperCamelCase_ ) with open(UpperCamelCase_ , 'rb' ) as f: SCREAMING_SNAKE_CASE__ = pickle.load(UpperCamelCase_ )['weights'] set_model_weights_in_torch(UpperCamelCase_ , UpperCamelCase_ , config.hidden_size ) # Save pytorch-model print(F'Save PyTorch model to {pytorch_dump_path}' ) torch.save(model.state_dict() , UpperCamelCase_ ) if __name__ == "__main__": __snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( """--trax_model_pkl_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained Reformer model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) __snake_case = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
169
0
'''simple docstring''' import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow UpperCamelCase__: Union[str, Any] = logging.getLogger() @unittest.skip("""Temporarily disable the doc tests.""" ) @require_torch @require_tf @slow class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : Optional[int] , __snake_case : Path , __snake_case : Union[str, None] = None , __snake_case : Union[List[str], None] = None , __snake_case : Union[str, List[str], None] = None , __snake_case : bool = True , ) -> Tuple: UpperCAmelCase : Optional[Any] = [file for file in os.listdir(__snake_case ) if os.path.isfile(os.path.join(__snake_case , __snake_case ) )] if identifier is not None: UpperCAmelCase : str = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(__snake_case , __snake_case ): for n_ in n_identifier: UpperCAmelCase : Dict = [file for file in files if n_ not in file] else: UpperCAmelCase : Dict = [file for file in files if n_identifier not in file] UpperCAmelCase : int = ignore_files or [] ignore_files.append('''__init__.py''' ) UpperCAmelCase : List[str] = [file for file in files if file not in ignore_files] for file in files: # Open all files print('''Testing''' , __snake_case ) if only_modules: UpperCAmelCase : List[str] = file.split('''.''' )[0] try: UpperCAmelCase : Dict = getattr(__snake_case , __snake_case ) UpperCAmelCase : Optional[int] = doctest.DocTestSuite(__snake_case ) UpperCAmelCase : str = unittest.TextTestRunner().run(__snake_case ) self.assertIs(len(result.failures ) , 0 ) except AttributeError: logger.info(F"""{module_identifier} is not a module.""" ) else: UpperCAmelCase : Any = doctest.testfile(str('''..''' / directory / file ) , optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed , 0 ) def A ( self : Any ) -> Tuple: UpperCAmelCase : List[str] = Path('''src/transformers''' ) UpperCAmelCase : int = '''modeling''' UpperCAmelCase : Tuple = [ '''modeling_ctrl.py''', '''modeling_tf_ctrl.py''', ] self.analyze_directory(__snake_case , identifier=__snake_case , ignore_files=__snake_case ) def A ( self : List[Any] ) -> Optional[Any]: UpperCAmelCase : List[Any] = Path('''src/transformers''' ) UpperCAmelCase : Tuple = '''tokenization''' self.analyze_directory(__snake_case , identifier=__snake_case ) def A ( self : Optional[int] ) -> Optional[Any]: UpperCAmelCase : Union[str, Any] = Path('''src/transformers''' ) UpperCAmelCase : str = '''configuration''' self.analyze_directory(__snake_case , identifier=__snake_case ) def A ( self : int ) -> Dict: UpperCAmelCase : Union[str, Any] = Path('''src/transformers''' ) UpperCAmelCase : str = ['''configuration''', '''modeling''', '''tokenization'''] self.analyze_directory(__snake_case , n_identifier=__snake_case ) def A ( self : Optional[int] ) -> Optional[Any]: UpperCAmelCase : Dict = Path('''docs/source''' ) UpperCAmelCase : Union[str, Any] = ['''favicon.ico'''] self.analyze_directory(__snake_case , ignore_files=__snake_case , only_modules=__snake_case )
23
import math import random def A__ ( __lowerCamelCase, __lowerCamelCase = False ): if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value __UpperCAmelCase = 0.02 def A__ ( __lowerCamelCase, __lowerCamelCase ): SCREAMING_SNAKE_CASE_ = float(2 * (random.randint(1, 1_00 )) - 1 ) for _ in range(__lowerCamelCase ): # Forward propagation SCREAMING_SNAKE_CASE_ = sigmoid_function(INITIAL_VALUE * weight ) # How much did we miss? SCREAMING_SNAKE_CASE_ = (expected / 1_00) - layer_a # Error delta SCREAMING_SNAKE_CASE_ = layer_1_error * sigmoid_function(__lowerCamelCase, __lowerCamelCase ) # Update weight weight += INITIAL_VALUE * layer_1_delta return layer_a * 1_00 if __name__ == "__main__": import doctest doctest.testmod() __UpperCAmelCase = int(input("Expected value: ")) __UpperCAmelCase = int(input("Number of propagations: ")) print(forward_propagation(expected, number_propagations))
299
0
"""simple docstring""" import unittest from transformers import MPNetConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) class lowercase_ : '''simple docstring''' def __init__( self : List[Any] , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : List[str]=13 , _UpperCAmelCase : Tuple=7 , _UpperCAmelCase : str=True , _UpperCAmelCase : List[str]=True , _UpperCAmelCase : Union[str, Any]=False , _UpperCAmelCase : int=True , _UpperCAmelCase : Optional[Any]=99 , _UpperCAmelCase : Any=64 , _UpperCAmelCase : Dict=5 , _UpperCAmelCase : List[str]=4 , _UpperCAmelCase : Union[str, Any]=64 , _UpperCAmelCase : Union[str, Any]="gelu" , _UpperCAmelCase : Any=0.1 , _UpperCAmelCase : Optional[Any]=0.1 , _UpperCAmelCase : Any=512 , _UpperCAmelCase : Tuple=16 , _UpperCAmelCase : Any=2 , _UpperCAmelCase : Optional[Any]=0.02 , _UpperCAmelCase : str=3 , _UpperCAmelCase : str=4 , _UpperCAmelCase : str=None , ): _A = parent _A = batch_size _A = seq_length _A = is_training _A = use_input_mask _A = use_token_type_ids _A = use_labels _A = vocab_size _A = hidden_size _A = num_hidden_layers _A = num_attention_heads _A = intermediate_size _A = hidden_act _A = hidden_dropout_prob _A = attention_probs_dropout_prob _A = max_position_embeddings _A = type_vocab_size _A = type_sequence_label_size _A = initializer_range _A = num_labels _A = num_choices _A = scope def lowerCAmelCase_ ( self : Dict ): return MPNetConfig.from_pretrained('microsoft/mpnet-base' ) def lowerCAmelCase_ ( self : Dict ): _A = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _A = None if self.use_input_mask: _A = random_attention_mask([self.batch_size, self.seq_length] ) _A = None _A = None _A = None if self.use_labels: _A = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _A = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _A = ids_tensor([self.batch_size] , self.num_choices ) _A = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCAmelCase_ ( self : Dict ): return MPNetConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def lowerCAmelCase_ ( self : Optional[Any] , _UpperCAmelCase : int , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : str , _UpperCAmelCase : Tuple , _UpperCAmelCase : Optional[Any] ): _A = MPNetModel(config=_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() _A = model(_UpperCAmelCase , _UpperCAmelCase ) _A = model(_UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def lowerCAmelCase_ ( self : Any , _UpperCAmelCase : Dict , _UpperCAmelCase : List[str] , _UpperCAmelCase : Any , _UpperCAmelCase : int , _UpperCAmelCase : Dict , _UpperCAmelCase : int ): _A = MPNetForQuestionAnswering(config=_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() _A = model( _UpperCAmelCase , attention_mask=_UpperCAmelCase , start_positions=_UpperCAmelCase , end_positions=_UpperCAmelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCAmelCase_ ( self : Tuple , _UpperCAmelCase : List[str] , _UpperCAmelCase : str , _UpperCAmelCase : str , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : List[Any] , _UpperCAmelCase : Dict ): _A = self.num_labels _A = MPNetForSequenceClassification(_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() _A = model(_UpperCAmelCase , attention_mask=_UpperCAmelCase , labels=_UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCAmelCase_ ( self : List[str] , _UpperCAmelCase : str , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : int , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : int , _UpperCAmelCase : int ): _A = self.num_choices _A = MPNetForMultipleChoice(config=_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() _A = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _A = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _A = model( _UpperCAmelCase , attention_mask=_UpperCAmelCase , labels=_UpperCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCAmelCase_ ( self : str , _UpperCAmelCase : Tuple , _UpperCAmelCase : List[str] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : List[str] , _UpperCAmelCase : List[str] , _UpperCAmelCase : List[str] ): _A = self.num_labels _A = MPNetForTokenClassification(config=_UpperCAmelCase ) model.to(_UpperCAmelCase ) model.eval() _A = model(_UpperCAmelCase , attention_mask=_UpperCAmelCase , labels=_UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCAmelCase_ ( self : Tuple ): _A = self.prepare_config_and_inputs() ((_A) , (_A) , (_A) , (_A) , (_A) , (_A)) = config_and_inputs _A = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class lowercase_ ( __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase : Tuple = ( ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) if is_torch_available() else () ) UpperCAmelCase : Union[str, Any] = ( { '''feature-extraction''': MPNetModel, '''fill-mask''': MPNetForMaskedLM, '''question-answering''': MPNetForQuestionAnswering, '''text-classification''': MPNetForSequenceClassification, '''token-classification''': MPNetForTokenClassification, '''zero-shot''': MPNetForSequenceClassification, } if is_torch_available() else {} ) UpperCAmelCase : Dict = False UpperCAmelCase : List[str] = True def lowerCAmelCase_ ( self : List[Any] ): _A = MPNetModelTester(self ) _A = ConfigTester(self , config_class=_UpperCAmelCase , hidden_size=37 ) def lowerCAmelCase_ ( self : int ): self.config_tester.run_common_tests() def lowerCAmelCase_ ( self : Union[str, Any] ): _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_model(*_UpperCAmelCase ) def lowerCAmelCase_ ( self : Optional[int] ): _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_sequence_classification(*_UpperCAmelCase ) def lowerCAmelCase_ ( self : Optional[Any] ): _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_multiple_choice(*_UpperCAmelCase ) def lowerCAmelCase_ ( self : List[Any] ): _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_token_classification(*_UpperCAmelCase ) def lowerCAmelCase_ ( self : List[str] ): _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_question_answering(*_UpperCAmelCase ) @require_torch class lowercase_ ( unittest.TestCase ): '''simple docstring''' @slow def lowerCAmelCase_ ( self : int ): _A = MPNetModel.from_pretrained('microsoft/mpnet-base' ) _A = torch.tensor([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) _A = model(_UpperCAmelCase )[0] _A = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , _UpperCAmelCase ) _A = torch.tensor( [[[-0.0550, 0.1943, -0.0740], [-0.0562, 0.2211, -0.0579], [-0.0437, 0.3337, -0.0641]]] ) # compare the actual values for a slice. self.assertTrue(torch.allclose(output[:, :3, :3] , _UpperCAmelCase , atol=1E-4 ) )
271
"""simple docstring""" from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) a = logging.get_logger(__name__) # pylint: disable=invalid-name a = ''' Examples: ```py >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline >>> import torch >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained("kandinsky-community/kandinsky-2-2-prior") >>> pipe_prior.to("cuda") >>> prompt = "red cat, 4k photo" >>> out = pipe_prior(prompt) >>> image_emb = out.image_embeds >>> zero_image_emb = out.negative_image_embeds >>> pipe = KandinskyV22Pipeline.from_pretrained("kandinsky-community/kandinsky-2-2-decoder") >>> pipe.to("cuda") >>> image = pipe( ... image_embeds=image_emb, ... negative_image_embeds=zero_image_emb, ... height=768, ... width=768, ... num_inference_steps=50, ... ).images >>> image[0].save("cat.png") ``` ''' def _snake_case ( _snake_case : Dict , _snake_case : List[Any] , _snake_case : List[Any]=8 ) -> Optional[int]: '''simple docstring''' _A = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 _A = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class lowercase_ ( __lowerCAmelCase ): '''simple docstring''' def __init__( self : Dict , _UpperCAmelCase : UNetaDConditionModel , _UpperCAmelCase : DDPMScheduler , _UpperCAmelCase : VQModel , ): super().__init__() self.register_modules( unet=_UpperCAmelCase , scheduler=_UpperCAmelCase , movq=_UpperCAmelCase , ) _A = 2 ** (len(self.movq.config.block_out_channels ) - 1) def lowerCAmelCase_ ( self : Any , _UpperCAmelCase : str , _UpperCAmelCase : Tuple , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Dict , _UpperCAmelCase : List[str] , _UpperCAmelCase : int ): if latents is None: _A = randn_tensor(_UpperCAmelCase , generator=_UpperCAmelCase , device=_UpperCAmelCase , dtype=_UpperCAmelCase ) else: if latents.shape != shape: raise ValueError(F'''Unexpected latents shape, got {latents.shape}, expected {shape}''' ) _A = latents.to(_UpperCAmelCase ) _A = latents * scheduler.init_noise_sigma return latents def lowerCAmelCase_ ( self : Tuple , _UpperCAmelCase : Tuple=0 ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) _A = torch.device(F'''cuda:{gpu_id}''' ) _A = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(_UpperCAmelCase , _UpperCAmelCase ) def lowerCAmelCase_ ( self : Union[str, Any] , _UpperCAmelCase : Union[str, Any]=0 ): if is_accelerate_available() and is_accelerate_version('>=' , '0.17.0.dev0' ): from accelerate import cpu_offload_with_hook else: raise ImportError('`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.' ) _A = torch.device(F'''cuda:{gpu_id}''' ) if self.device.type != "cpu": self.to('cpu' , silence_dtype_warnings=_UpperCAmelCase ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) _A = None for cpu_offloaded_model in [self.unet, self.movq]: _A , _A = cpu_offload_with_hook(_UpperCAmelCase , _UpperCAmelCase , prev_module_hook=_UpperCAmelCase ) # We'll offload the last model manually. _A = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def lowerCAmelCase_ ( self : int ): if not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(_UpperCAmelCase , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(_UpperCAmelCase ) def __call__( self : Union[str, Any] , _UpperCAmelCase : Union[torch.FloatTensor, List[torch.FloatTensor]] , _UpperCAmelCase : Union[torch.FloatTensor, List[torch.FloatTensor]] , _UpperCAmelCase : int = 512 , _UpperCAmelCase : int = 512 , _UpperCAmelCase : int = 100 , _UpperCAmelCase : float = 4.0 , _UpperCAmelCase : int = 1 , _UpperCAmelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , _UpperCAmelCase : Optional[torch.FloatTensor] = None , _UpperCAmelCase : Optional[str] = "pil" , _UpperCAmelCase : bool = True , ): _A = self._execution_device _A = guidance_scale > 1.0 if isinstance(_UpperCAmelCase , _UpperCAmelCase ): _A = torch.cat(_UpperCAmelCase , dim=0 ) _A = image_embeds.shape[0] * num_images_per_prompt if isinstance(_UpperCAmelCase , _UpperCAmelCase ): _A = torch.cat(_UpperCAmelCase , dim=0 ) if do_classifier_free_guidance: _A = image_embeds.repeat_interleave(_UpperCAmelCase , dim=0 ) _A = negative_image_embeds.repeat_interleave(_UpperCAmelCase , dim=0 ) _A = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=_UpperCAmelCase ) self.scheduler.set_timesteps(_UpperCAmelCase , device=_UpperCAmelCase ) _A = self.scheduler.timesteps _A = self.unet.config.in_channels _A , _A = downscale_height_and_width(_UpperCAmelCase , _UpperCAmelCase , self.movq_scale_factor ) # create initial latent _A = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , self.scheduler , ) for i, t in enumerate(self.progress_bar(_UpperCAmelCase ) ): # expand the latents if we are doing classifier free guidance _A = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents _A = {'image_embeds': image_embeds} _A = self.unet( sample=_UpperCAmelCase , timestep=_UpperCAmelCase , encoder_hidden_states=_UpperCAmelCase , added_cond_kwargs=_UpperCAmelCase , return_dict=_UpperCAmelCase , )[0] if do_classifier_free_guidance: _A , _A = noise_pred.split(latents.shape[1] , dim=1 ) _A , _A = noise_pred.chunk(2 ) _A , _A = variance_pred.chunk(2 ) _A = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) _A = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , 'variance_type' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): _A , _A = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 _A = self.scheduler.step( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , generator=_UpperCAmelCase , )[0] # post-processing _A = self.movq.decode(_UpperCAmelCase , force_not_quantize=_UpperCAmelCase )['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(F'''Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}''' ) if output_type in ["np", "pil"]: _A = image * 0.5 + 0.5 _A = image.clamp(0 , 1 ) _A = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": _A = self.numpy_to_pil(_UpperCAmelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=_UpperCAmelCase )
271
1
import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing the experiment tracking capability, # and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To help focus on the differences in the code, building `DataLoaders` # was refactored into its own function. # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## _UpperCAmelCase : List[Any] = 16 _UpperCAmelCase : Any = 32 def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase = 16 ) -> List[str]: lowerCamelCase__ : Any = AutoTokenizer.from_pretrained('bert-base-cased' ) lowerCamelCase__ : List[Any] = load_dataset('glue' , 'mrpc' ) def tokenize_function(_UpperCAmelCase ): # max_length=None => use the model max length (it's actually the default) lowerCamelCase__ : Optional[int] = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=_UpperCAmelCase , max_length=_UpperCAmelCase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): lowerCamelCase__ : List[Any] = datasets.map( _UpperCAmelCase , batched=_UpperCAmelCase , remove_columns=['idx', 'sentence1', 'sentence2'] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library lowerCamelCase__ : List[Any] = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(_UpperCAmelCase ): # On TPU it's best to pad everything to the same length or training will be very slow. lowerCamelCase__ : int = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": lowerCamelCase__ : int = 16 elif accelerator.mixed_precision != "no": lowerCamelCase__ : Dict = 8 else: lowerCamelCase__ : Optional[int] = None return tokenizer.pad( _UpperCAmelCase , padding='longest' , max_length=_UpperCAmelCase , pad_to_multiple_of=_UpperCAmelCase , return_tensors='pt' , ) # Instantiate dataloaders. lowerCamelCase__ : Optional[int] = DataLoader( tokenized_datasets['train'] , shuffle=_UpperCAmelCase , collate_fn=_UpperCAmelCase , batch_size=_UpperCAmelCase ) lowerCamelCase__ : str = DataLoader( tokenized_datasets['validation'] , shuffle=_UpperCAmelCase , collate_fn=_UpperCAmelCase , batch_size=_UpperCAmelCase ) return train_dataloader, eval_dataloader # For testing only if os.environ.get("""TESTING_MOCKED_DATALOADERS""", None) == "1": from accelerate.test_utils.training import mocked_dataloaders _UpperCAmelCase : Optional[Any] = mocked_dataloaders # noqa: F811 def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> str: # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS' , _UpperCAmelCase ) == "1": lowerCamelCase__ : Optional[int] = 2 # Initialize Accelerator # New Code # # We pass in "all" to `log_with` to grab all available trackers in the environment # Note: If using a custom `Tracker` class, should be passed in here such as: # >>> log_with = ["all", MyCustomTrackerClassInstance()] if args.with_tracking: lowerCamelCase__ : List[Any] = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , log_with='all' , project_dir=args.project_dir ) else: lowerCamelCase__ : Optional[int] = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs lowerCamelCase__ : Optional[Any] = config['lr'] lowerCamelCase__ : Tuple = int(config['num_epochs'] ) lowerCamelCase__ : Union[str, Any] = int(config['seed'] ) lowerCamelCase__ : str = int(config['batch_size'] ) set_seed(_UpperCAmelCase ) lowerCamelCase__ , lowerCamelCase__ : Optional[int] = get_dataloaders(_UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase__ : Optional[int] = evaluate.load('glue' , 'mrpc' ) # If the batch size is too big we use gradient accumulation lowerCamelCase__ : Optional[Any] = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: lowerCamelCase__ : Any = batch_size // MAX_GPU_BATCH_SIZE lowerCamelCase__ : str = MAX_GPU_BATCH_SIZE # Instantiate the model (we build the model here so that the seed also control new weights initialization) lowerCamelCase__ : Union[str, Any] = AutoModelForSequenceClassification.from_pretrained('bert-base-cased' , return_dict=_UpperCAmelCase ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). lowerCamelCase__ : Optional[Any] = model.to(accelerator.device ) # Instantiate optimizer lowerCamelCase__ : List[str] = AdamW(params=model.parameters() , lr=_UpperCAmelCase ) # Instantiate scheduler lowerCamelCase__ : List[Any] = get_linear_schedule_with_warmup( optimizer=_UpperCAmelCase , num_warmup_steps=100 , num_training_steps=(len(_UpperCAmelCase ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = accelerator.prepare( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # New Code # # We need to initialize the trackers we use. Overall configurations can also be stored if args.with_tracking: lowerCamelCase__ : Optional[Any] = os.path.split(_UpperCAmelCase )[-1].split('.' )[0] accelerator.init_trackers(_UpperCAmelCase , _UpperCAmelCase ) # Now we train the model for epoch in range(_UpperCAmelCase ): model.train() # New Code # # For our tracking example, we will log the total loss of each epoch if args.with_tracking: lowerCamelCase__ : Union[str, Any] = 0 for step, batch in enumerate(_UpperCAmelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) lowerCamelCase__ : Any = model(**_UpperCAmelCase ) lowerCamelCase__ : Dict = outputs.loss # New Code # if args.with_tracking: total_loss += loss.detach().float() lowerCamelCase__ : Dict = loss / gradient_accumulation_steps accelerator.backward(_UpperCAmelCase ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(_UpperCAmelCase ): # We could avoid this line since we set the accelerator with `device_placement=True` (the default). batch.to(accelerator.device ) with torch.no_grad(): lowerCamelCase__ : str = model(**_UpperCAmelCase ) lowerCamelCase__ : List[str] = outputs.logits.argmax(dim=-1 ) lowerCamelCase__ , lowerCamelCase__ : Dict = accelerator.gather_for_metrics((predictions, batch['labels']) ) metric.add_batch( predictions=_UpperCAmelCase , references=_UpperCAmelCase , ) lowerCamelCase__ : Tuple = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F"""epoch {epoch}:""" , _UpperCAmelCase ) # New Code # # To actually log, we call `Accelerator.log` # The values passed can be of `str`, `int`, `float` or `dict` of `str` to `float`/`int` if args.with_tracking: accelerator.log( { 'accuracy': eval_metric['accuracy'], 'f1': eval_metric['f1'], 'train_loss': total_loss.item() / len(_UpperCAmelCase ), 'epoch': epoch, } , step=_UpperCAmelCase , ) # New Code # # When a run is finished, you should call `accelerator.end_training()` # to close all of the open trackers if args.with_tracking: accelerator.end_training() def SCREAMING_SNAKE_CASE ( ) -> List[str]: lowerCamelCase__ : Any = argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument( '--mixed_precision' , type=_UpperCAmelCase , default=_UpperCAmelCase , choices=['no', 'fp16', 'bf16', 'fp8'] , help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' , ) parser.add_argument('--cpu' , action='store_true' , help='If passed, will train on the CPU.' ) parser.add_argument( '--with_tracking' , action='store_true' , help='Whether to load in all available experiment trackers from the environment and use them for logging.' , ) parser.add_argument( '--project_dir' , type=_UpperCAmelCase , default='logs' , help='Location on where to store experiment tracking logs` and relevent project information' , ) lowerCamelCase__ : Union[str, Any] = parser.parse_args() lowerCamelCase__ : str = {'lr': 2e-5, 'num_epochs': 3, 'seed': 42, 'batch_size': 16} training_function(_UpperCAmelCase , _UpperCAmelCase ) if __name__ == "__main__": main()
50
'''simple docstring''' from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": A =input('Enter image url: ').strip() print(f"""Downloading image from {url} ...""") A =BeautifulSoup(requests.get(url).content, 'html.parser') # The image URL is in the content field of the first meta tag with property og:image A =soup.find('meta', {'property': 'og:image'})['content'] A =requests.get(image_url).content A =f"""{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg""" with open(file_name, 'wb') as fp: fp.write(image_data) print(f"""Done. Image saved to disk as {file_name}.""")
34
0
'''simple docstring''' import logging import numpy as np import pytest from scipy.linalg import eigh logging.basicConfig(level=logging.INFO, format='''%(message)s''') def _SCREAMING_SNAKE_CASE ( UpperCamelCase ): """simple docstring""" return input_array.reshape((input_array.size, 1) ) def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase ): """simple docstring""" lowerCAmelCase__ : List[Any] = np.nan for i in range(UpperCamelCase ): lowerCAmelCase__ : List[Any] = features[:, labels == i] lowerCAmelCase__ : List[Any] = data.mean(1 ) # Centralize the data of class i lowerCAmelCase__ : Union[str, Any] = data - column_reshape(UpperCamelCase ) if i > 0: # If covariance_sum is not None covariance_sum += np.dot(UpperCamelCase , centered_data.T ) else: # If covariance_sum is np.nan (i.e. first loop) lowerCAmelCase__ : Union[str, Any] = np.dot(UpperCamelCase , centered_data.T ) return covariance_sum / features.shape[1] def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase ): """simple docstring""" lowerCAmelCase__ : str = features.mean(1 ) lowerCAmelCase__ : Dict = np.nan for i in range(UpperCamelCase ): lowerCAmelCase__ : str = features[:, labels == i] lowerCAmelCase__ : List[str] = data.shape[1] lowerCAmelCase__ : Tuple = data.mean(1 ) if i > 0: # If covariance_sum is not None covariance_sum += device_data * np.dot( column_reshape(UpperCamelCase ) - column_reshape(UpperCamelCase ) , (column_reshape(UpperCamelCase ) - column_reshape(UpperCamelCase )).T , ) else: # If covariance_sum is np.nan (i.e. first loop) lowerCAmelCase__ : Dict = device_data * np.dot( column_reshape(UpperCamelCase ) - column_reshape(UpperCamelCase ) , (column_reshape(UpperCamelCase ) - column_reshape(UpperCamelCase )).T , ) return covariance_sum / features.shape[1] def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase ): """simple docstring""" if features.any(): lowerCAmelCase__ : Union[str, Any] = features.mean(1 ) # Center the dataset lowerCAmelCase__ : Union[str, Any] = features - np.reshape(UpperCamelCase , (data_mean.size, 1) ) lowerCAmelCase__ : str = np.dot(UpperCamelCase , centered_data.T ) / features.shape[1] lowerCAmelCase__ , lowerCAmelCase__ : int = np.linalg.eigh(UpperCamelCase ) # Take all the columns in the reverse order (-1), and then takes only the first lowerCAmelCase__ : Union[str, Any] = eigenvectors[:, ::-1][:, 0:dimensions] # Project the database on the new space lowerCAmelCase__ : Any = np.dot(filtered_eigenvectors.T , UpperCamelCase ) logging.info("""Principal Component Analysis computed""" ) return projected_data else: logging.basicConfig(level=logging.ERROR , format="""%(message)s""" , force=UpperCamelCase ) logging.error("""Dataset empty""" ) raise AssertionError def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ): """simple docstring""" assert classes > dimensions # Check if features have been already loaded if features.any: lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = eigh( covariance_between_classes(UpperCamelCase , UpperCamelCase , UpperCamelCase ) , covariance_within_classes(UpperCamelCase , UpperCamelCase , UpperCamelCase ) , ) lowerCAmelCase__ : Optional[Any] = eigenvectors[:, ::-1][:, :dimensions] lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = np.linalg.svd(UpperCamelCase ) lowerCAmelCase__ : Optional[Any] = svd_matrix[:, 0:dimensions] lowerCAmelCase__ : Any = np.dot(filtered_svd_matrix.T , UpperCamelCase ) logging.info("""Linear Discriminant Analysis computed""" ) return projected_data else: logging.basicConfig(level=logging.ERROR , format="""%(message)s""" , force=UpperCamelCase ) logging.error("""Dataset empty""" ) raise AssertionError def _SCREAMING_SNAKE_CASE ( ): """simple docstring""" lowerCAmelCase__ : Union[str, Any] = np.array([[1, 2, 3, 4, 5], [2, 3, 4, 5, 6], [3, 4, 5, 6, 7]] ) lowerCAmelCase__ : int = np.array([0, 0, 0, 1, 1] ) lowerCAmelCase__ : str = 2 lowerCAmelCase__ : Tuple = 2 # Assert that the function raises an AssertionError if dimensions > classes with pytest.raises(UpperCamelCase ) as error_info: lowerCAmelCase__ : Optional[Any] = linear_discriminant_analysis( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) if isinstance(UpperCamelCase , np.ndarray ): raise AssertionError( """Did not raise AssertionError for dimensions > classes""" ) assert error_info.type is AssertionError def _SCREAMING_SNAKE_CASE ( ): """simple docstring""" lowerCAmelCase__ : int = np.array([[1, 2, 3], [4, 5, 6], [7, 8, 9]] ) lowerCAmelCase__ : str = 2 lowerCAmelCase__ : Any = np.array([[6.9282_0323, 8.6602_5404, 10.3923_0485], [3.0, 3.0, 3.0]] ) with pytest.raises(UpperCamelCase ) as error_info: lowerCAmelCase__ : Any = principal_component_analysis(UpperCamelCase , UpperCamelCase ) if not np.allclose(UpperCamelCase , UpperCamelCase ): raise AssertionError assert error_info.type is AssertionError if __name__ == "__main__": import doctest doctest.testmod()
184
'''simple docstring''' # Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def _SCREAMING_SNAKE_CASE ( UpperCamelCase ): """simple docstring""" return 1 / (1 + np.exp(-z )) def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase ): """simple docstring""" return (-y * np.log(UpperCamelCase ) - (1 - y) * np.log(1 - h )).mean() def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase ): """simple docstring""" lowerCAmelCase__ : Union[str, Any] = np.dot(UpperCamelCase , UpperCamelCase ) return np.sum(y * scores - np.log(1 + np.exp(UpperCamelCase ) ) ) def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase=70000 ): """simple docstring""" lowerCAmelCase__ : Union[str, Any] = np.zeros(x.shape[1] ) for iterations in range(UpperCamelCase ): lowerCAmelCase__ : str = np.dot(UpperCamelCase , UpperCamelCase ) lowerCAmelCase__ : Tuple = sigmoid_function(UpperCamelCase ) lowerCAmelCase__ : List[str] = np.dot(x.T , h - y ) / y.size lowerCAmelCase__ : List[Any] = theta - alpha * gradient # updating the weights lowerCAmelCase__ : List[Any] = np.dot(UpperCamelCase , UpperCamelCase ) lowerCAmelCase__ : List[Any] = sigmoid_function(UpperCamelCase ) lowerCAmelCase__ : List[Any] = cost_function(UpperCamelCase , UpperCamelCase ) if iterations % 100 == 0: print(f"""loss: {j} \t""" ) # printing the loss after every 100 iterations return theta # In[68]: if __name__ == "__main__": _lowerCAmelCase = datasets.load_iris() _lowerCAmelCase = iris.data[:, :2] _lowerCAmelCase = (iris.target != 0) * 1 _lowerCAmelCase = 0.1 _lowerCAmelCase = logistic_reg(alpha, x, y, max_iterations=7_0000) print('''theta: ''', theta) # printing the theta i.e our weights vector def _SCREAMING_SNAKE_CASE ( UpperCamelCase ): """simple docstring""" return sigmoid_function( np.dot(UpperCamelCase , UpperCamelCase ) ) # predicting the value of probability from the logistic regression algorithm plt.figure(figsize=(10, 6)) plt.scatter(x[y == 0][:, 0], x[y == 0][:, 1], color='''b''', label='''0''') plt.scatter(x[y == 1][:, 0], x[y == 1][:, 1], color='''r''', label='''1''') ((_lowerCAmelCase) , (_lowerCAmelCase)) = (x[:, 0].min(), x[:, 0].max()) ((_lowerCAmelCase) , (_lowerCAmelCase)) = (x[:, 1].min(), x[:, 1].max()) ((_lowerCAmelCase) , (_lowerCAmelCase)) = np.meshgrid(np.linspace(xa_min, xa_max), np.linspace(xa_min, xa_max)) _lowerCAmelCase = np.c_[xxa.ravel(), xxa.ravel()] _lowerCAmelCase = predict_prob(grid).reshape(xxa.shape) plt.contour(xxa, xxa, probs, [0.5], linewidths=1, colors='''black''') plt.legend() plt.show()
184
1
'''simple docstring''' import torch from diffusers import KDPMaDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class UpperCAmelCase_ ( __lowercase ): lowerCamelCase : str = (KDPMaDiscreteScheduler,) lowerCamelCase : Optional[int] = 10 def __UpperCAmelCase ( self : List[Any] , **UpperCAmelCase__ : Optional[int] ) -> Any: lowerCAmelCase = { 'num_train_timesteps': 1_1_0_0, 'beta_start': 0.0_001, 'beta_end': 0.02, 'beta_schedule': 'linear', } config.update(**UpperCAmelCase__ ) return config def __UpperCAmelCase ( self : List[str] ) -> Tuple: for timesteps in [1_0, 5_0, 1_0_0, 1_0_0_0]: self.check_over_configs(num_train_timesteps=UpperCAmelCase__ ) def __UpperCAmelCase ( self : Dict ) -> Optional[Any]: for beta_start, beta_end in zip([0.00_001, 0.0_001, 0.001] , [0.0_002, 0.002, 0.02] ): self.check_over_configs(beta_start=UpperCAmelCase__ , beta_end=UpperCAmelCase__ ) def __UpperCAmelCase ( self : Optional[Any] ) -> int: for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=UpperCAmelCase__ ) def __UpperCAmelCase ( self : Union[str, Any] ) -> List[str]: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=UpperCAmelCase__ ) def __UpperCAmelCase ( self : int ) -> Dict: lowerCAmelCase = self.scheduler_classes[0] lowerCAmelCase = self.get_scheduler_config(prediction_type='v_prediction' ) lowerCAmelCase = scheduler_class(**UpperCAmelCase__ ) scheduler.set_timesteps(self.num_inference_steps ) lowerCAmelCase = self.dummy_model() lowerCAmelCase = self.dummy_sample_deter * scheduler.init_noise_sigma lowerCAmelCase = sample.to(UpperCAmelCase__ ) for i, t in enumerate(scheduler.timesteps ): lowerCAmelCase = scheduler.scale_model_input(UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = model(UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = output.prev_sample lowerCAmelCase = torch.sum(torch.abs(UpperCAmelCase__ ) ) lowerCAmelCase = torch.mean(torch.abs(UpperCAmelCase__ ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 4.69_34E-07 ) < 1E-2 assert abs(result_mean.item() - 6.11_12E-10 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 4.6_93_42_86_50_17_09_72E-07 ) < 1E-2 assert abs(result_mean.item() - 0.0_002 ) < 1E-3 def __UpperCAmelCase ( self : Tuple ) -> Dict: if torch_device == "mps": return lowerCAmelCase = self.scheduler_classes[0] lowerCAmelCase = self.get_scheduler_config() lowerCAmelCase = scheduler_class(**UpperCAmelCase__ ) scheduler.set_timesteps(self.num_inference_steps ) lowerCAmelCase = self.dummy_model() lowerCAmelCase = self.dummy_sample_deter * scheduler.init_noise_sigma lowerCAmelCase = sample.to(UpperCAmelCase__ ) for i, t in enumerate(scheduler.timesteps ): lowerCAmelCase = scheduler.scale_model_input(UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = model(UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = output.prev_sample lowerCAmelCase = torch.sum(torch.abs(UpperCAmelCase__ ) ) lowerCAmelCase = torch.mean(torch.abs(UpperCAmelCase__ ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 20.4_125 ) < 1E-2 assert abs(result_mean.item() - 0.0_266 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 20.4_125 ) < 1E-2 assert abs(result_mean.item() - 0.0_266 ) < 1E-3 def __UpperCAmelCase ( self : List[Any] ) -> Optional[Any]: if torch_device == "mps": return lowerCAmelCase = self.scheduler_classes[0] lowerCAmelCase = self.get_scheduler_config() lowerCAmelCase = scheduler_class(**UpperCAmelCase__ ) scheduler.set_timesteps(self.num_inference_steps , device=UpperCAmelCase__ ) lowerCAmelCase = self.dummy_model() lowerCAmelCase = self.dummy_sample_deter.to(UpperCAmelCase__ ) * scheduler.init_noise_sigma for t in scheduler.timesteps: lowerCAmelCase = scheduler.scale_model_input(UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = model(UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase = output.prev_sample lowerCAmelCase = torch.sum(torch.abs(UpperCAmelCase__ ) ) lowerCAmelCase = torch.mean(torch.abs(UpperCAmelCase__ ) ) if str(UpperCAmelCase__ ).startswith('cpu' ): # The following sum varies between 148 and 156 on mps. Why? assert abs(result_sum.item() - 20.4_125 ) < 1E-2 assert abs(result_mean.item() - 0.0_266 ) < 1E-3 else: # CUDA assert abs(result_sum.item() - 20.4_125 ) < 1E-2 assert abs(result_mean.item() - 0.0_266 ) < 1E-3
4
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case =logging.get_logger(__name__) __snake_case ={ """facebook/s2t-wav2vec2-large-en-de""": ( """https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/config.json""" ), # See all Speech2Text models at https://huggingface.co/models?filter=speech2text2 } class UpperCAmelCase_ ( __lowercase ): lowerCamelCase : Union[str, Any] = '''speech_to_text_2''' lowerCamelCase : Any = ['''past_key_values'''] lowerCamelCase : Optional[Any] = {'''num_attention_heads''': '''decoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : Optional[int] , UpperCAmelCase__ : Optional[Any]=1_0_0_0_0 , UpperCAmelCase__ : int=6 , UpperCAmelCase__ : Optional[Any]=2_0_4_8 , UpperCAmelCase__ : Union[str, Any]=4 , UpperCAmelCase__ : List[Any]=0.0 , UpperCAmelCase__ : str=True , UpperCAmelCase__ : str="relu" , UpperCAmelCase__ : Any=2_5_6 , UpperCAmelCase__ : Optional[Any]=0.1 , UpperCAmelCase__ : Any=0.0 , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : List[Any]=0.02 , UpperCAmelCase__ : List[Any]=2 , UpperCAmelCase__ : str=True , UpperCAmelCase__ : List[str]=1 , UpperCAmelCase__ : Any=0 , UpperCAmelCase__ : Dict=2 , UpperCAmelCase__ : int=1_0_2_4 , **UpperCAmelCase__ : Optional[Any] , ) -> Dict: lowerCAmelCase = vocab_size lowerCAmelCase = d_model lowerCAmelCase = decoder_ffn_dim lowerCAmelCase = decoder_layers lowerCAmelCase = decoder_attention_heads lowerCAmelCase = dropout lowerCAmelCase = attention_dropout lowerCAmelCase = activation_dropout lowerCAmelCase = activation_function lowerCAmelCase = init_std lowerCAmelCase = decoder_layerdrop lowerCAmelCase = use_cache lowerCAmelCase = decoder_layers lowerCAmelCase = scale_embedding # scale factor will be sqrt(d_model) if True lowerCAmelCase = max_target_positions super().__init__( pad_token_id=UpperCAmelCase__ , bos_token_id=UpperCAmelCase__ , eos_token_id=UpperCAmelCase__ , decoder_start_token_id=UpperCAmelCase__ , **UpperCAmelCase__ , )
4
1
"""simple docstring""" import collections import os from typing import List, Optional, Tuple from transformers.utils import is_jieba_available, requires_backends if is_jieba_available(): import jieba from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __lowerCAmelCase : List[Any] =logging.get_logger(__name__) __lowerCAmelCase : Optional[int] ={"""vocab_file""": """vocab.txt"""} __lowerCAmelCase : Tuple ={ """vocab_file""": { """openbmb/cpm-ant-10b""": """https://huggingface.co/openbmb/cpm-ant-10b/blob/main/vocab.txt""", }, } __lowerCAmelCase : List[Any] ={ """openbmb/cpm-ant-10b""": 1_0_2_4, } def UpperCAmelCase__ ( lowerCAmelCase__ :List[str] ) -> Optional[int]: '''simple docstring''' lowercase = collections.OrderedDict() with open(lowerCAmelCase__ , """r""" , encoding="""utf-8""" ) as reader: lowercase = reader.readlines() for index, token in enumerate(lowerCAmelCase__ ): lowercase = token.rstrip("""\n""" ) lowercase = index return vocab class _A ( lowerCAmelCase ): def __init__( self , __lowerCAmelCase , __lowerCAmelCase="<unk>" , __lowerCAmelCase=200 ): """simple docstring""" lowercase = vocab lowercase = unk_token lowercase = max_input_chars_per_word def A__ ( self , __lowerCAmelCase ): """simple docstring""" lowercase = list(__lowerCAmelCase ) if len(__lowerCAmelCase ) > self.max_input_chars_per_word: return [self.unk_token] lowercase = 0 lowercase = [] while start < len(__lowerCAmelCase ): lowercase = len(__lowerCAmelCase ) lowercase = None while start < end: lowercase = """""".join(chars[start:end] ) if substr in self.vocab: lowercase = substr break end -= 1 if cur_substr is None: sub_tokens.append(self.unk_token ) start += 1 else: sub_tokens.append(__lowerCAmelCase ) lowercase = end return sub_tokens class _A ( lowerCAmelCase ): snake_case__ : Optional[int] = VOCAB_FILES_NAMES snake_case__ : Optional[int] = PRETRAINED_VOCAB_FILES_MAP snake_case__ : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case__ : Optional[Any] = ['input_ids', 'attention_mask'] snake_case__ : Dict = False def __init__( self , __lowerCAmelCase , __lowerCAmelCase="<d>" , __lowerCAmelCase="</d>" , __lowerCAmelCase="<s>" , __lowerCAmelCase="</s>" , __lowerCAmelCase="<pad>" , __lowerCAmelCase="<unk>" , __lowerCAmelCase="</n>" , __lowerCAmelCase="</_>" , __lowerCAmelCase="left" , **__lowerCAmelCase , ): """simple docstring""" requires_backends(self , ["""jieba"""] ) super().__init__( bod_token=__lowerCAmelCase , eod_token=__lowerCAmelCase , bos_token=__lowerCAmelCase , eos_token=__lowerCAmelCase , pad_token=__lowerCAmelCase , unk_token=__lowerCAmelCase , line_token=__lowerCAmelCase , space_token=__lowerCAmelCase , padding_side=__lowerCAmelCase , **__lowerCAmelCase , ) lowercase = bod_token lowercase = eod_token lowercase = load_vocab(__lowerCAmelCase ) lowercase = self.encoder[space_token] lowercase = self.encoder[line_token] del self.encoder[space_token] del self.encoder[line_token] lowercase = collections.OrderedDict(sorted(self.encoder.items() , key=lambda __lowerCAmelCase : x[1] ) ) lowercase = {v: k for k, v in self.encoder.items()} lowercase = WordpieceTokenizer(vocab=self.encoder , unk_token=self.unk_token ) @property def A__ ( self ): """simple docstring""" return self.encoder[self.bod_token] @property def A__ ( self ): """simple docstring""" return self.encoder[self.eod_token] @property def A__ ( self ): """simple docstring""" return self.encoder["\n"] @property def A__ ( self ): """simple docstring""" return len(self.encoder ) def A__ ( self ): """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def A__ ( self , __lowerCAmelCase ): """simple docstring""" lowercase = [] for x in jieba.cut(__lowerCAmelCase , cut_all=__lowerCAmelCase ): output_tokens.extend(self.wordpiece_tokenizer.tokenize(__lowerCAmelCase ) ) return output_tokens def A__ ( self , __lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" lowercase = [i for i in token_ids if i >= 0] lowercase = [ x for x in token_ids if x != self.pad_token_id and x != self.eos_token_id and x != self.bos_token_id ] return super()._decode(__lowerCAmelCase , **__lowerCAmelCase ) def A__ ( self , __lowerCAmelCase ): """simple docstring""" return token in self.encoder def A__ ( self , __lowerCAmelCase ): """simple docstring""" return "".join(__lowerCAmelCase ) def A__ ( self , __lowerCAmelCase ): """simple docstring""" return self.encoder.get(__lowerCAmelCase , self.encoder.get(self.unk_token ) ) def A__ ( self , __lowerCAmelCase ): """simple docstring""" return self.decoder.get(__lowerCAmelCase , self.unk_token ) def A__ ( self , __lowerCAmelCase , __lowerCAmelCase = None ): """simple docstring""" if os.path.isdir(__lowerCAmelCase ): lowercase = os.path.join( __lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) else: lowercase = (filename_prefix + """-""" if filename_prefix else """""") + save_directory lowercase = 0 if " " in self.encoder: lowercase = self.encoder[""" """] del self.encoder[" "] if "\n" in self.encoder: lowercase = self.encoder["""\n"""] del self.encoder["\n"] lowercase = collections.OrderedDict(sorted(self.encoder.items() , key=lambda __lowerCAmelCase : x[1] ) ) with open(__lowerCAmelCase , """w""" , encoding="""utf-8""" ) as writer: for token, token_index in self.encoder.items(): if index != token_index: logger.warning( f'Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.' """ Please check that the vocabulary is not corrupted!""" ) lowercase = token_index writer.write(token + """\n""" ) index += 1 return (vocab_file,) def A__ ( self , __lowerCAmelCase , __lowerCAmelCase = None ): """simple docstring""" if token_ids_a is None: return [self.bos_token_id] + token_ids_a return [self.bos_token_id] + token_ids_a + [self.bos_token_id] + token_ids_a def A__ ( self , __lowerCAmelCase , __lowerCAmelCase = None , __lowerCAmelCase = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__lowerCAmelCase , token_ids_a=__lowerCAmelCase , already_has_special_tokens=__lowerCAmelCase ) if token_ids_a is not None: return [1] + ([0] * len(__lowerCAmelCase )) + [1] + ([0] * len(__lowerCAmelCase )) return [1] + ([0] * len(__lowerCAmelCase ))
32
"""simple docstring""" def UpperCAmelCase__ ( lowerCAmelCase__ :list[list] ) -> list[list]: '''simple docstring''' lowercase = current_set.copy() for row_index, row in enumerate(lowerCAmelCase__ ): lowercase = row[0] for column_index, column in enumerate(lowerCAmelCase__ ): if magnitude == 0: lowercase = column continue lowercase = column / magnitude # Subtract to cancel term lowercase = current_set[0] lowercase = [first_row] lowercase = current_set[1::] for row in current_set: lowercase = [] # If first term is 0, it is already in form we want, so we preserve it if row[0] == 0: final_set.append(lowerCAmelCase__ ) continue for column_index in range(len(lowerCAmelCase__ ) ): temp_row.append(first_row[column_index] - row[column_index] ) final_set.append(lowerCAmelCase__ ) # Create next recursion iteration set if len(final_set[0] ) != 3: lowercase = final_set[0] lowercase = [] lowercase = [] for row in final_set[1::]: current_first_column.append(row[0] ) next_iteration.append(row[1::] ) lowercase = simplify(lowerCAmelCase__ ) for i in range(len(lowerCAmelCase__ ) ): resultant[i].insert(0 , current_first_column[i] ) resultant.insert(0 , lowerCAmelCase__ ) lowercase = resultant return final_set def UpperCAmelCase__ ( lowerCAmelCase__ :list[list] ) -> list: '''simple docstring''' if len(lowerCAmelCase__ ) == 0: raise IndexError("""solve_simultaneous() requires n lists of length n+1""" ) lowercase = len(lowerCAmelCase__ ) + 1 if any(len(lowerCAmelCase__ ) != _length for item in equations ): raise IndexError("""solve_simultaneous() requires n lists of length n+1""" ) for row in equations: if any(not isinstance(lowerCAmelCase__ , (int, float) ) for column in row ): raise ValueError("""solve_simultaneous() requires lists of integers""" ) if len(lowerCAmelCase__ ) == 1: return [equations[0][-1] / equations[0][0]] lowercase = equations.copy() if any(0 in row for row in data_set ): lowercase = data_set.copy() lowercase = [] for row_index, row in enumerate(lowerCAmelCase__ ): if 0 not in row: lowercase = data_set.pop(lowerCAmelCase__ ) break if not full_row: raise ValueError("""solve_simultaneous() requires at least 1 full equation""" ) data_set.insert(0 , lowerCAmelCase__ ) lowercase = data_set.copy() lowercase = simplify(lowerCAmelCase__ ) lowercase = simplified[::-1] lowercase = [] for row in simplified: lowercase = row[-1] if not solutions: if row[-2] == 0: solutions.append(0 ) continue solutions.append(current_solution / row[-2] ) continue lowercase = row.copy()[: len(lowerCAmelCase__ ) - 1 :] while temp_row[0] == 0: temp_row.pop(0 ) if len(lowerCAmelCase__ ) == 0: solutions.append(0 ) continue lowercase = temp_row[1::] lowercase = temp_row[::-1] for column_index, column in enumerate(lowerCAmelCase__ ): current_solution -= column * solutions[column_index] solutions.append(lowerCAmelCase__ ) lowercase = [] for item in solutions: final.append(float(round(lowerCAmelCase__ , 5 ) ) ) return final[::-1] if __name__ == "__main__": import doctest doctest.testmod() __lowerCAmelCase : List[str] =[ [2, 1, 1, 1, 1, 4], [1, 2, 1, 1, 1, 5], [1, 1, 2, 1, 1, 6], [1, 1, 1, 2, 1, 7], [1, 1, 1, 1, 2, 8], ] print(solve_simultaneous(eq)) print(solve_simultaneous([[4, 2]]))
32
1
'''simple docstring''' import argparse import os import torch from transformers import ( XLNetConfig, XLNetForQuestionAnswering, XLNetForSequenceClassification, XLNetLMHeadModel, load_tf_weights_in_xlnet, ) from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging __lowerCAmelCase = { '''cola''': 2, '''mnli''': 3, '''mrpc''': 2, '''sst-2''': 2, '''sts-b''': 1, '''qqp''': 2, '''qnli''': 2, '''rte''': 2, '''wnli''': 2, } logging.set_verbosity_info() def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None ) -> Dict: # Initialise PyTorch model _a : Union[str, Any] = XLNetConfig.from_json_file(lowerCAmelCase_ ) _a : Any = finetuning_task.lower() if finetuning_task is not None else '' if finetuning_task in GLUE_TASKS_NUM_LABELS: print(f"""Building PyTorch XLNetForSequenceClassification model from configuration: {config}""" ) _a : List[Any] = finetuning_task _a : str = GLUE_TASKS_NUM_LABELS[finetuning_task] _a : Dict = XLNetForSequenceClassification(lowerCAmelCase_ ) elif "squad" in finetuning_task: _a : int = finetuning_task _a : List[Any] = XLNetForQuestionAnswering(lowerCAmelCase_ ) else: _a : Dict = XLNetLMHeadModel(lowerCAmelCase_ ) # Load weights from tf checkpoint load_tf_weights_in_xlnet(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # Save pytorch-model _a : Tuple = os.path.join(lowerCAmelCase_ , lowerCAmelCase_ ) _a : str = os.path.join(lowerCAmelCase_ , lowerCAmelCase_ ) print(f"""Save PyTorch model to {os.path.abspath(lowerCAmelCase_ )}""" ) torch.save(model.state_dict() , lowerCAmelCase_ ) print(f"""Save configuration file to {os.path.abspath(lowerCAmelCase_ )}""" ) with open(lowerCAmelCase_ , 'w' , encoding='utf-8' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--xlnet_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained XLNet model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the folder to store the PyTorch model or dataset/vocab.''', ) parser.add_argument( '''--finetuning_task''', default=None, type=str, help='''Name of a task on which the XLNet TensorFlow model was fine-tuned''', ) __lowerCAmelCase = parser.parse_args() print(args) convert_xlnet_checkpoint_to_pytorch( args.tf_checkpoint_path, args.xlnet_config_file, args.pytorch_dump_folder_path, args.finetuning_task )
89
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _SCREAMING_SNAKE_CASE : int = { "configuration_clip": [ "CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "CLIPConfig", "CLIPOnnxConfig", "CLIPTextConfig", "CLIPVisionConfig", ], "processing_clip": ["CLIPProcessor"], "tokenization_clip": ["CLIPTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE : Optional[int] = ["CLIPTokenizerFast"] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE : Optional[Any] = ["CLIPFeatureExtractor"] _SCREAMING_SNAKE_CASE : Dict = ["CLIPImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE : List[Any] = [ "CLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "CLIPModel", "CLIPPreTrainedModel", "CLIPTextModel", "CLIPTextModelWithProjection", "CLIPVisionModel", "CLIPVisionModelWithProjection", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE : List[str] = [ "TF_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "TFCLIPModel", "TFCLIPPreTrainedModel", "TFCLIPTextModel", "TFCLIPVisionModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE : Any = [ "FlaxCLIPModel", "FlaxCLIPPreTrainedModel", "FlaxCLIPTextModel", "FlaxCLIPTextPreTrainedModel", "FlaxCLIPVisionModel", "FlaxCLIPVisionPreTrainedModel", ] if TYPE_CHECKING: from .configuration_clip import ( CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, CLIPConfig, CLIPOnnxConfig, CLIPTextConfig, CLIPVisionConfig, ) from .processing_clip import CLIPProcessor from .tokenization_clip import CLIPTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_clip_fast import CLIPTokenizerFast try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_clip import CLIPFeatureExtractor from .image_processing_clip import CLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_clip import ( CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, CLIPModel, CLIPPreTrainedModel, CLIPTextModel, CLIPTextModelWithProjection, CLIPVisionModel, CLIPVisionModelWithProjection, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_clip import ( TF_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFCLIPModel, TFCLIPPreTrainedModel, TFCLIPTextModel, TFCLIPVisionModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_clip import ( FlaxCLIPModel, FlaxCLIPPreTrainedModel, FlaxCLIPTextModel, FlaxCLIPTextPreTrainedModel, FlaxCLIPVisionModel, FlaxCLIPVisionPreTrainedModel, ) else: import sys _SCREAMING_SNAKE_CASE : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
85
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _UpperCamelCase: Dict = {'configuration_wavlm': ['WAVLM_PRETRAINED_CONFIG_ARCHIVE_MAP', 'WavLMConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase: Optional[int] = [ 'WAVLM_PRETRAINED_MODEL_ARCHIVE_LIST', 'WavLMForAudioFrameClassification', 'WavLMForCTC', 'WavLMForSequenceClassification', 'WavLMForXVector', 'WavLMModel', 'WavLMPreTrainedModel', ] if TYPE_CHECKING: from .configuration_wavlm import WAVLM_PRETRAINED_CONFIG_ARCHIVE_MAP, WavLMConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_wavlm import ( WAVLM_PRETRAINED_MODEL_ARCHIVE_LIST, WavLMForAudioFrameClassification, WavLMForCTC, WavLMForSequenceClassification, WavLMForXVector, WavLMModel, WavLMPreTrainedModel, ) else: import sys _UpperCamelCase: List[str] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
367
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class a__ : def __init__( self : Union[str, Any], lowerCAmelCase : Any, lowerCAmelCase : Tuple=13, lowerCAmelCase : List[Any]=2, lowerCAmelCase : Tuple=24, lowerCAmelCase : Any=16, lowerCAmelCase : Optional[Any]=True, lowerCAmelCase : Tuple=True, lowerCAmelCase : Optional[int]=32, lowerCAmelCase : Optional[int]=5, lowerCAmelCase : Optional[int]=4, lowerCAmelCase : Optional[int]=37, lowerCAmelCase : Tuple="gelu", lowerCAmelCase : str=0.1, lowerCAmelCase : Tuple=0.1, lowerCAmelCase : List[Any]=10, lowerCAmelCase : List[Any]=0.02, lowerCAmelCase : List[str]=None, lowerCAmelCase : Any=2, lowerCAmelCase : str=2, ) -> Union[str, Any]: lowercase : str = parent lowercase : Optional[int] = batch_size lowercase : str = patch_size lowercase : List[Any] = max_length lowercase : Optional[Any] = num_mel_bins lowercase : int = is_training lowercase : Dict = use_labels lowercase : List[str] = hidden_size lowercase : str = num_hidden_layers lowercase : Any = num_attention_heads lowercase : List[str] = intermediate_size lowercase : int = hidden_act lowercase : Optional[Any] = hidden_dropout_prob lowercase : Optional[Any] = attention_probs_dropout_prob lowercase : int = type_sequence_label_size lowercase : Optional[int] = initializer_range lowercase : int = scope lowercase : int = frequency_stride lowercase : Dict = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) lowercase : Tuple = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 lowercase : Dict = (self.max_length - self.patch_size) // self.time_stride + 1 lowercase : Any = frequency_out_dimension * time_out_dimension lowercase : List[str] = num_patches + 2 def lowercase ( self : int ) -> Optional[int]: lowercase : List[Any] = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) lowercase : List[Any] = None if self.use_labels: lowercase : Optional[Any] = ids_tensor([self.batch_size], self.type_sequence_label_size ) lowercase : str = self.get_config() return config, input_values, labels def lowercase ( self : List[str] ) -> Any: return ASTConfig( patch_size=self.patch_size, max_length=self.max_length, num_mel_bins=self.num_mel_bins, hidden_size=self.hidden_size, num_hidden_layers=self.num_hidden_layers, num_attention_heads=self.num_attention_heads, intermediate_size=self.intermediate_size, hidden_act=self.hidden_act, hidden_dropout_prob=self.hidden_dropout_prob, attention_probs_dropout_prob=self.attention_probs_dropout_prob, is_decoder=lowerCAmelCase, initializer_range=self.initializer_range, frequency_stride=self.frequency_stride, time_stride=self.time_stride, ) def lowercase ( self : str, lowerCAmelCase : List[Any], lowerCAmelCase : Optional[Any], lowerCAmelCase : Union[str, Any] ) -> Optional[int]: lowercase : Any = ASTModel(config=lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() lowercase : Any = model(lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase ( self : Any ) -> Tuple: lowercase : List[Any] = self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ) : Dict = config_and_inputs lowercase : Union[str, Any] = {'input_values': input_values} return config, inputs_dict @require_torch class a__ ( SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__, unittest.TestCase ): _lowerCamelCase = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) _lowerCamelCase = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) _lowerCamelCase = False _lowerCamelCase = False _lowerCamelCase = False _lowerCamelCase = False def lowercase ( self : Any, lowerCAmelCase : Any, lowerCAmelCase : Tuple, lowerCAmelCase : Dict, lowerCAmelCase : List[str], lowerCAmelCase : int ) -> Tuple: if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def lowercase ( self : Optional[Any] ) -> Dict: lowercase : List[Any] = ASTModelTester(self ) lowercase : Any = ConfigTester(self, config_class=lowerCAmelCase, has_text_modality=lowerCAmelCase, hidden_size=37 ) def lowercase ( self : Tuple ) -> Optional[int]: self.config_tester.run_common_tests() @unittest.skip(reason='AST does not use inputs_embeds' ) def lowercase ( self : Tuple ) -> List[Any]: pass def lowercase ( self : Union[str, Any] ) -> List[str]: lowercase , lowercase : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase : Optional[Any] = model_class(lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings(), (nn.Module) ) lowercase : int = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCAmelCase, nn.Linear ) ) def lowercase ( self : Union[str, Any] ) -> Optional[Any]: lowercase , lowercase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase : Optional[int] = model_class(lowerCAmelCase ) lowercase : Optional[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase : List[Any] = [*signature.parameters.keys()] lowercase : str = ['input_values'] self.assertListEqual(arg_names[:1], lowerCAmelCase ) def lowercase ( self : Optional[int] ) -> Tuple: lowercase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase ) @slow def lowercase ( self : List[str] ) -> Optional[Any]: for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase : Dict = ASTModel.from_pretrained(lowerCAmelCase ) self.assertIsNotNone(lowerCAmelCase ) def lowercase__ ( ) -> Any: '''simple docstring''' lowercase : Tuple = hf_hub_download( repo_id='nielsr/audio-spectogram-transformer-checkpoint' , filename='sample_audio.flac' , repo_type='dataset' ) lowercase , lowercase : List[str] = torchaudio.load(_UpperCAmelCase ) return audio, sampling_rate @require_torch @require_torchaudio class a__ ( unittest.TestCase ): @cached_property def lowercase ( self : Union[str, Any] ) -> Optional[int]: return ( ASTFeatureExtractor.from_pretrained('MIT/ast-finetuned-audioset-10-10-0.4593' ) if is_torchaudio_available() else None ) @slow def lowercase ( self : Any ) -> Optional[Any]: lowercase : List[str] = self.default_feature_extractor lowercase : Tuple = ASTForAudioClassification.from_pretrained('MIT/ast-finetuned-audioset-10-10-0.4593' ).to(lowerCAmelCase ) lowercase : List[str] = self.default_feature_extractor lowercase , lowercase : Optional[int] = prepare_audio() lowercase : List[str] = audio.squeeze().numpy() lowercase : List[Any] = feature_extractor(lowerCAmelCase, sampling_rate=lowerCAmelCase, return_tensors='pt' ).to(lowerCAmelCase ) # forward pass with torch.no_grad(): lowercase : List[Any] = model(**lowerCAmelCase ) # verify the logits lowercase : Union[str, Any] = torch.Size((1, 527) ) self.assertEqual(outputs.logits.shape, lowerCAmelCase ) lowercase : Any = torch.tensor([-0.8760, -7.0042, -8.6602] ).to(lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3], lowerCAmelCase, atol=1e-4 ) )
53
0
'''simple docstring''' import inspect import unittest from transformers import ViTHybridConfig from transformers.testing_utils import require_accelerate, require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel from transformers.models.vit_hybrid.modeling_vit_hybrid import VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class UpperCamelCase__ : """simple docstring""" def __init__( self , snake_case , snake_case=1_3 , snake_case=6_4 , snake_case=2 , snake_case=3 , snake_case=True , snake_case=True , snake_case=3_2 , snake_case=5 , snake_case=4 , snake_case=3_7 , snake_case="gelu" , snake_case=0.1 , snake_case=0.1 , snake_case=1_0 , snake_case=0.02 , snake_case=[1, 1_6, 4, 4] , snake_case=None , ): '''simple docstring''' UpperCAmelCase : Optional[Any] = parent UpperCAmelCase : Union[str, Any] = batch_size UpperCAmelCase : Any = image_size UpperCAmelCase : List[str] = patch_size UpperCAmelCase : Dict = num_channels UpperCAmelCase : int = is_training UpperCAmelCase : List[str] = use_labels UpperCAmelCase : Optional[int] = hidden_size UpperCAmelCase : Tuple = num_hidden_layers UpperCAmelCase : Any = num_attention_heads UpperCAmelCase : Optional[int] = intermediate_size UpperCAmelCase : List[Any] = hidden_act UpperCAmelCase : Union[str, Any] = hidden_dropout_prob UpperCAmelCase : List[Any] = attention_probs_dropout_prob UpperCAmelCase : Optional[int] = type_sequence_label_size UpperCAmelCase : Optional[Any] = initializer_range UpperCAmelCase : Dict = scope UpperCAmelCase : Union[str, Any] = backbone_featmap_shape # in ViT hybrid, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) # the number of patches is based on the feature map of the backbone, which by default uses an output stride # of 32, which means that the feature map has a spatial resolution of 1/32 of the input image size UpperCAmelCase : List[str] = (self.image_size // 3_2) ** 2 UpperCAmelCase : Optional[Any] = num_patches + 1 def A_ ( self ): '''simple docstring''' UpperCAmelCase : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase : List[str] = None if self.use_labels: UpperCAmelCase : Dict = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase : Tuple = self.get_config() return config, pixel_values, labels def A_ ( self ): '''simple docstring''' UpperCAmelCase : Optional[int] = { "global_padding": "same", "layer_type": "bottleneck", "depths": [3, 4, 9], "out_features": ["stage1", "stage2", "stage3"], "embedding_dynamic_padding": True, "hidden_sizes": [4, 8, 1_6, 3_2], "num_groups": 2, } return ViTHybridConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowercase , initializer_range=self.initializer_range , backbone_featmap_shape=self.backbone_featmap_shape , backbone_config=_lowercase , ) def A_ ( self , snake_case , snake_case , snake_case ): '''simple docstring''' UpperCAmelCase : List[str] = ViTHybridModel(config=_lowercase ) model.to(_lowercase ) model.eval() UpperCAmelCase : Union[str, Any] = model(_lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A_ ( self , snake_case , snake_case , snake_case ): '''simple docstring''' UpperCAmelCase : str = self.type_sequence_label_size UpperCAmelCase : List[Any] = ViTHybridForImageClassification(_lowercase ) model.to(_lowercase ) model.eval() UpperCAmelCase : Union[str, Any] = model(_lowercase , labels=_lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def A_ ( self ): '''simple docstring''' UpperCAmelCase : Optional[int] = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : Union[str, Any] = config_and_inputs UpperCAmelCase : int = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class UpperCamelCase__ ( lowerCamelCase_ , lowerCamelCase_ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE__ : List[Any] = (ViTHybridModel, ViTHybridForImageClassification) if is_torch_available() else () SCREAMING_SNAKE_CASE__ : Tuple = ( {"feature-extraction": ViTHybridModel, "image-classification": ViTHybridForImageClassification} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE__ : Any = False SCREAMING_SNAKE_CASE__ : Optional[int] = False SCREAMING_SNAKE_CASE__ : int = False def A_ ( self ): '''simple docstring''' UpperCAmelCase : Optional[int] = ViTHybridModelTester(self ) UpperCAmelCase : Any = ConfigTester(self , config_class=_lowercase , has_text_modality=_lowercase , hidden_size=3_7 ) def A_ ( self ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="ViT does not use inputs_embeds" ) def A_ ( self ): '''simple docstring''' pass def A_ ( self ): '''simple docstring''' UpperCAmelCase , UpperCAmelCase : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase : Any = model_class(_lowercase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) UpperCAmelCase : List[str] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowercase , nn.Linear ) ) def A_ ( self ): '''simple docstring''' UpperCAmelCase , UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase : int = model_class(_lowercase ) UpperCAmelCase : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase : Optional[Any] = [*signature.parameters.keys()] UpperCAmelCase : str = ["pixel_values"] self.assertListEqual(arg_names[:1] , _lowercase ) def A_ ( self ): '''simple docstring''' UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowercase ) def A_ ( self ): '''simple docstring''' UpperCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowercase ) def A_ ( self ): '''simple docstring''' UpperCAmelCase , UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase : int = _config_zero_init(_lowercase ) for model_class in self.all_model_classes: UpperCAmelCase : Union[str, Any] = model_class(config=_lowercase ) # Skip the check for the backbone for name, module in model.named_modules(): if module.__class__.__name__ == "ViTHybridPatchEmbeddings": UpperCAmelCase : Optional[Any] = [f"{name}.{key}" for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=f"Parameter {name} of model {model_class} seems not properly initialized" , ) @slow def A_ ( self ): '''simple docstring''' for model_name in VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase : str = ViTHybridModel.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) def lowercase ( ): '''simple docstring''' UpperCAmelCase : Dict = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" @cached_property def A_ ( self ): '''simple docstring''' return ( ViTHybridImageProcessor.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def A_ ( self ): '''simple docstring''' UpperCAmelCase : int = ViTHybridForImageClassification.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to( _lowercase ) UpperCAmelCase : Any = self.default_image_processor UpperCAmelCase : Union[str, Any] = prepare_img() UpperCAmelCase : Tuple = image_processor(images=_lowercase , return_tensors="pt" ).to(_lowercase ) # forward pass with torch.no_grad(): UpperCAmelCase : Any = model(**_lowercase ) # verify the logits UpperCAmelCase : Tuple = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , _lowercase ) UpperCAmelCase : Optional[Any] = torch.tensor([-1.9090, -0.4993, -0.2389] ).to(_lowercase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowercase , atol=1e-4 ) ) @slow @require_accelerate def A_ ( self ): '''simple docstring''' UpperCAmelCase : Optional[Any] = ViTHybridImageProcessor.from_pretrained("google/vit-hybrid-base-bit-384" ) UpperCAmelCase : List[Any] = ViTHybridForImageClassification.from_pretrained("google/vit-hybrid-base-bit-384" , device_map="auto" ) UpperCAmelCase : List[str] = prepare_img() UpperCAmelCase : Any = image_processor(images=_lowercase , return_tensors="pt" ) UpperCAmelCase : Any = model(**_lowercase ) UpperCAmelCase : List[Any] = outputs.logits # model predicts one of the 1000 ImageNet classes UpperCAmelCase : Dict = logits.argmax(-1 ).item() self.assertTrue(model.config.idalabel[predicted_class_idx] , "tabby, tabby cat" )
311
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ..models.auto import AutoModelForVisionaSeq from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class __snake_case ( lowerCamelCase_ ): lowerCAmelCase_ = "Salesforce/blip-image-captioning-base" lowerCAmelCase_ = ( "This is a tool that generates a description of an image. It takes an input named `image` which should be the " "image to caption, and returns a text that contains the description in English." ) lowerCAmelCase_ = "image_captioner" lowerCAmelCase_ = AutoModelForVisionaSeq lowerCAmelCase_ = ["image"] lowerCAmelCase_ = ["text"] def __init__( self : List[Any] , *_lowercase : Optional[int] , **_lowercase : Union[str, Any] ): """simple docstring""" requires_backends(self , ["""vision"""] ) super().__init__(*_lowercase , **_lowercase ) def __a ( self : Tuple , _lowercase : "Image" ): """simple docstring""" return self.pre_processor(images=_lowercase , return_tensors="""pt""" ) def __a ( self : Union[str, Any] , _lowercase : Optional[int] ): """simple docstring""" return self.model.generate(**_lowercase ) def __a ( self : int , _lowercase : Any ): """simple docstring""" return self.pre_processor.batch_decode(_lowercase , skip_special_tokens=_lowercase )[0].strip()
219
0
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_barthez import BarthezTokenizer else: lowerCamelCase = None lowerCamelCase = logging.get_logger(__name__) lowerCamelCase = {"""vocab_file""": """sentencepiece.bpe.model""", """tokenizer_file""": """tokenizer.json"""} lowerCamelCase = { """vocab_file""": { """moussaKam/mbarthez""": """https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model""", """moussaKam/barthez""": """https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model""", """moussaKam/barthez-orangesum-title""": ( """https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model""" ), }, """tokenizer_file""": { """moussaKam/mbarthez""": """https://huggingface.co/moussaKam/mbarthez/resolve/main/tokenizer.json""", """moussaKam/barthez""": """https://huggingface.co/moussaKam/barthez/resolve/main/tokenizer.json""", """moussaKam/barthez-orangesum-title""": ( """https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/tokenizer.json""" ), }, } lowerCamelCase = { """moussaKam/mbarthez""": 1_024, """moussaKam/barthez""": 1_024, """moussaKam/barthez-orangesum-title""": 1_024, } lowerCamelCase = """▁""" class lowercase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' UpperCamelCase = VOCAB_FILES_NAMES UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase = ['''input_ids''', '''attention_mask'''] UpperCamelCase = BarthezTokenizer def __init__( self : str , _UpperCAmelCase : Union[str, Any]=None , _UpperCAmelCase : Optional[int]=None , _UpperCAmelCase : Optional[int]="<s>" , _UpperCAmelCase : Optional[int]="</s>" , _UpperCAmelCase : Optional[int]="</s>" , _UpperCAmelCase : List[str]="<s>" , _UpperCAmelCase : Tuple="<unk>" , _UpperCAmelCase : str="<pad>" , _UpperCAmelCase : List[Any]="<mask>" , **_UpperCAmelCase : Dict , ) -> List[Any]: '''simple docstring''' UpperCAmelCase_ = AddedToken(_UpperCAmelCase , lstrip=_UpperCAmelCase , rstrip=_UpperCAmelCase ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ) else mask_token super().__init__( _UpperCAmelCase , tokenizer_file=_UpperCAmelCase , bos_token=_UpperCAmelCase , eos_token=_UpperCAmelCase , unk_token=_UpperCAmelCase , sep_token=_UpperCAmelCase , cls_token=_UpperCAmelCase , pad_token=_UpperCAmelCase , mask_token=_UpperCAmelCase , **_UpperCAmelCase , ) UpperCAmelCase_ = vocab_file UpperCAmelCase_ = False if not self.vocab_file else True def lowercase__ ( self : Optional[Any] , _UpperCAmelCase : List[int] , _UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] UpperCAmelCase_ = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def lowercase__ ( self : Dict , _UpperCAmelCase : List[int] , _UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' UpperCAmelCase_ = [self.sep_token_id] UpperCAmelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowercase__ ( self : Union[str, Any] , _UpperCAmelCase : str , _UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(_UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase_ = os.path.join( _UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_UpperCAmelCase ): copyfile(self.vocab_file , _UpperCAmelCase ) return (out_vocab_file,)
369
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow if is_torch_available(): import torch from transformers import XLMRobertaModel @require_sentencepiece @require_tokenizers @require_torch class lowercase__ ( unittest.TestCase ): '''simple docstring''' @slow def lowercase__ ( self : Tuple ) -> str: '''simple docstring''' UpperCAmelCase_ = XLMRobertaModel.from_pretrained("xlm-roberta-base" ) UpperCAmelCase_ = torch.tensor([[0, 581, 10269, 83, 99942, 136, 60742, 23, 70, 80583, 18276, 2]] ) # The dog is cute and lives in the garden house UpperCAmelCase_ = torch.Size((1, 12, 768) ) # batch_size, sequence_length, embedding_vector_dim UpperCAmelCase_ = torch.tensor( [[-0.0101, 0.1218, -0.0803, 0.0801, 0.1327, 0.0776, -0.1215, 0.2383, 0.3338, 0.3106, 0.0300, 0.0252]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): UpperCAmelCase_ = model(_UpperCAmelCase )["last_hidden_state"].detach() self.assertEqual(output.shape , _UpperCAmelCase ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , _UpperCAmelCase , atol=1e-3 ) ) @slow def lowercase__ ( self : List[Any] ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ = XLMRobertaModel.from_pretrained("xlm-roberta-large" ) UpperCAmelCase_ = torch.tensor([[0, 581, 10269, 83, 99942, 136, 60742, 23, 70, 80583, 18276, 2]] ) # The dog is cute and lives in the garden house UpperCAmelCase_ = torch.Size((1, 12, 1024) ) # batch_size, sequence_length, embedding_vector_dim UpperCAmelCase_ = torch.tensor( [[-0.0699, -0.0318, 0.0705, -0.1241, 0.0999, -0.0520, 0.1004, -0.1838, -0.4704, 0.1437, 0.0821, 0.0126]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.large') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): UpperCAmelCase_ = model(_UpperCAmelCase )["last_hidden_state"].detach() self.assertEqual(output.shape , _UpperCAmelCase ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , _UpperCAmelCase , atol=1e-3 ) )
241
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ : Tuple = logging.get_logger(__name__) lowercase__ : List[str] = { 'unc-nlp/lxmert-base-uncased': 'https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json', } class __lowerCAmelCase ( __magic_name__ ): """simple docstring""" _snake_case : Tuple = 'lxmert' _snake_case : Optional[Any] = {} def __init__( self : Any , lowerCAmelCase__ : str=30522 , lowerCAmelCase__ : List[str]=768 , lowerCAmelCase__ : Optional[Any]=12 , lowerCAmelCase__ : Any=9500 , lowerCAmelCase__ : Union[str, Any]=1600 , lowerCAmelCase__ : List[Any]=400 , lowerCAmelCase__ : str=3072 , lowerCAmelCase__ : str="gelu" , lowerCAmelCase__ : Optional[int]=0.1 , lowerCAmelCase__ : Union[str, Any]=0.1 , lowerCAmelCase__ : Any=512 , lowerCAmelCase__ : Tuple=2 , lowerCAmelCase__ : int=0.02 , lowerCAmelCase__ : Any=1e-1_2 , lowerCAmelCase__ : Dict=9 , lowerCAmelCase__ : str=5 , lowerCAmelCase__ : Dict=5 , lowerCAmelCase__ : Any=2048 , lowerCAmelCase__ : Tuple=4 , lowerCAmelCase__ : Union[str, Any]=6.67 , lowerCAmelCase__ : List[Any]=True , lowerCAmelCase__ : List[str]=True , lowerCAmelCase__ : Tuple=True , lowerCAmelCase__ : List[Any]=True , lowerCAmelCase__ : Union[str, Any]=True , lowerCAmelCase__ : Optional[Any]=True , lowerCAmelCase__ : int=True , **lowerCAmelCase__ : List[str] , ) -> int: '''simple docstring''' _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_attention_heads _UpperCamelCase = hidden_act _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = type_vocab_size _UpperCamelCase = initializer_range _UpperCamelCase = layer_norm_eps _UpperCamelCase = num_qa_labels _UpperCamelCase = num_object_labels _UpperCamelCase = num_attr_labels _UpperCamelCase = l_layers _UpperCamelCase = x_layers _UpperCamelCase = r_layers _UpperCamelCase = visual_feat_dim _UpperCamelCase = visual_pos_dim _UpperCamelCase = visual_loss_normalizer _UpperCamelCase = task_matched _UpperCamelCase = task_mask_lm _UpperCamelCase = task_obj_predict _UpperCamelCase = task_qa _UpperCamelCase = visual_obj_loss _UpperCamelCase = visual_attr_loss _UpperCamelCase = visual_feat_loss _UpperCamelCase = {'''vision''': r_layers, '''cross_encoder''': x_layers, '''language''': l_layers} super().__init__(**lowerCAmelCase__ )
324
'''simple docstring''' import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() lowercase__ : Optional[int] = logging.get_logger(__name__) lowercase__ : str = [ ['attention', 'attn'], ['encoder_attention', 'encoder_attn'], ['q_lin', 'q_proj'], ['k_lin', 'k_proj'], ['v_lin', 'v_proj'], ['out_lin', 'out_proj'], ['norm_embeddings', 'layernorm_embedding'], ['position_embeddings', 'embed_positions'], ['embeddings', 'embed_tokens'], ['ffn.lin', 'fc'], ] def a__ ( lowercase : str ) -> Dict: """simple docstring""" if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: _UpperCamelCase = k.replace(lowercase, lowercase ) if k.startswith('''encoder''' ): _UpperCamelCase = k.replace('''.attn''', '''.self_attn''' ) _UpperCamelCase = k.replace('''norm1''', '''self_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm2''', '''final_layer_norm''' ) elif k.startswith('''decoder''' ): _UpperCamelCase = k.replace('''norm1''', '''self_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm2''', '''encoder_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm3''', '''final_layer_norm''' ) return k def a__ ( lowercase : List[str] ) -> List[Any]: """simple docstring""" _UpperCamelCase = [ '''model.encoder.layernorm_embedding.weight''', '''model.encoder.layernorm_embedding.bias''', '''model.decoder.layernorm_embedding.weight''', '''model.decoder.layernorm_embedding.bias''', ] for k in keys: _UpperCamelCase = sd.pop(lowercase ) _UpperCamelCase = k.replace('''layernorm_embedding''', '''layer_norm''' ) assert new_k not in sd _UpperCamelCase = v lowercase__ : str = ['START'] @torch.no_grad() def a__ ( lowercase : Optional[int], lowercase : List[str], lowercase : List[str] ) -> Dict: """simple docstring""" _UpperCamelCase = torch.load(lowercase, map_location='''cpu''' ) _UpperCamelCase = model['''model'''] _UpperCamelCase = BlenderbotConfig.from_json_file(lowercase ) _UpperCamelCase = BlenderbotForConditionalGeneration(lowercase ) _UpperCamelCase = m.model.state_dict().keys() _UpperCamelCase = [] _UpperCamelCase = {} for k, v in sd.items(): if k in IGNORE_KEYS: continue _UpperCamelCase = rename_state_dict_key(lowercase ) if new_k not in valid_keys: failures.append([k, new_k] ) else: _UpperCamelCase = v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(lowercase ) m.model.load_state_dict(lowercase, strict=lowercase ) m.half() m.save_pretrained(lowercase ) if __name__ == "__main__": lowercase__ : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument('--src_path', type=str, help='like blenderbot-model.bin') parser.add_argument('--save_dir', default='hf_blenderbot', type=str, help='Where to save converted model.') parser.add_argument( '--hf_config_json', default='blenderbot-3b-config.json', type=str, help='Path to config to use' ) lowercase__ : Optional[Any] = parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
324
1
"""simple docstring""" import argparse import os from pathlib import Path from typing import Dict import tensorflow as tf import torch from tqdm import tqdm from transformers import PegasusConfig, PegasusForConditionalGeneration, PegasusTokenizer from transformers.models.pegasus.configuration_pegasus import DEFAULTS, task_specific_params UpperCamelCase_ =[ # replace left string with right string to get the relevant state_dict key (identical state dict to bart) ["memory_attention", "encoder_attn"], ["attention", "attn"], ["/", "."], [".LayerNorm.gamma", "_layer_norm.weight"], [".LayerNorm.beta", "_layer_norm.bias"], ["r.layer_", "r.layers."], ["output_proj", "out_proj"], ["ffn.dense_1.", "fc2."], ["ffn.dense.", "fc1."], ["ffn_layer_norm", "final_layer_norm"], ["kernel", "weight"], ["encoder_layer_norm.", "encoder.layer_norm."], ["decoder_layer_norm.", "decoder.layer_norm."], ["embeddings.weights", "shared.weight"], ] def a_ ( _lowercase ): for pegasus_name, hf_name in PATTERNS: _UpperCamelCase : Optional[Any] = k.replace(lowerCamelCase__ , lowerCamelCase__ ) return k def a_ ( _lowercase , _lowercase ): _UpperCamelCase : List[str] = DEFAULTS.copy() cfg_kwargs.update(lowerCamelCase__ ) _UpperCamelCase : str = PegasusConfig(**lowerCamelCase__ ) _UpperCamelCase : List[str] = PegasusForConditionalGeneration(lowerCamelCase__ ) _UpperCamelCase : Optional[Any] = torch_model.model.state_dict() _UpperCamelCase : int = {} for k, v in tf_weights.items(): _UpperCamelCase : Optional[Any] = rename_state_dict_key(lowerCamelCase__ ) if new_k not in sd: raise ValueError(F"""could not find new key {new_k} in state dict. (converted from {k})""" ) if "dense" in k or "proj" in new_k: _UpperCamelCase : Optional[Any] = v.T _UpperCamelCase : Union[str, Any] = torch.tensor(lowerCamelCase__ , dtype=sd[new_k].dtype ) assert v.shape == sd[new_k].shape, F"""{new_k}, {k}, {v.shape}, {sd[new_k].shape}""" # make sure embedding.padding_idx is respected _UpperCamelCase : int = torch.zeros_like(mapping['''shared.weight'''][cfg.pad_token_id + 1] ) _UpperCamelCase : Optional[Any] = mapping['''shared.weight'''] _UpperCamelCase : List[Any] = mapping['''shared.weight'''] _UpperCamelCase : Dict = {k: torch.zeros_like(lowerCamelCase__ ) for k, v in sd.items() if k.endswith('''bias''' ) and k not in mapping} mapping.update(**lowerCamelCase__ ) _UpperCamelCase , _UpperCamelCase : Optional[Any] = torch_model.model.load_state_dict(lowerCamelCase__ , strict=lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = [ k for k in missing if k not in ['''encoder.embed_positions.weight''', '''decoder.embed_positions.weight'''] ] assert unexpected_missing == [], F"""no matches found for the following torch keys {unexpected_missing}""" assert extra == [], F"""no matches found for the following tf keys {extra}""" return torch_model def a_ ( _lowercase="./ckpt/aeslc/model.ckpt-32000" ): _UpperCamelCase : str = tf.train.list_variables(lowerCamelCase__ ) _UpperCamelCase : Tuple = {} _UpperCamelCase : Union[str, Any] = ['''Adafactor''', '''global_step'''] for name, shape in tqdm(lowerCamelCase__ , desc='''converting tf checkpoint to dict''' ): _UpperCamelCase : Tuple = any(pat in name for pat in ignore_name ) if skip_key: continue _UpperCamelCase : List[Any] = tf.train.load_variable(lowerCamelCase__ , lowerCamelCase__ ) _UpperCamelCase : List[str] = array return tf_weights def a_ ( _lowercase , _lowercase ): _UpperCamelCase : Dict = Path(lowerCamelCase__ ).parent.name _UpperCamelCase : Dict = task_specific_params[F"""summarization_{dataset}"""]['''max_position_embeddings'''] _UpperCamelCase : str = PegasusTokenizer.from_pretrained('''sshleifer/pegasus''' , model_max_length=lowerCamelCase__ ) assert tok.model_max_length == desired_max_model_length tok.save_pretrained(lowerCamelCase__ ) # convert model _UpperCamelCase : str = get_tf_weights_as_numpy(lowerCamelCase__ ) _UpperCamelCase : str = task_specific_params[F"""summarization_{dataset}"""] if dataset == "large": _UpperCamelCase : List[str] = task_specific_params _UpperCamelCase : int = convert_pegasus(lowerCamelCase__ , lowerCamelCase__ ) torch_model.save_pretrained(lowerCamelCase__ ) _UpperCamelCase : Tuple = torch_model.state_dict() sd.pop('''model.decoder.embed_positions.weight''' ) sd.pop('''model.encoder.embed_positions.weight''' ) torch.save(lowerCamelCase__ , Path(lowerCamelCase__ ) / '''pytorch_model.bin''' ) if __name__ == "__main__": UpperCamelCase_ =argparse.ArgumentParser() # Required parameters parser.add_argument("""tf_ckpt_path""", type=str, help="""passed to tf.train.list_variables""") parser.add_argument("""save_dir""", default=None, type=str, help="""Path to the output PyTorch model.""") UpperCamelCase_ =parser.parse_args() if args.save_dir is None: UpperCamelCase_ =Path(args.tf_ckpt_path).parent.name UpperCamelCase_ =os.path.join("""pegasus""", dataset) convert_pegasus_ckpt_to_pytorch(args.tf_ckpt_path, args.save_dir)
350
"""simple docstring""" import argparse import json from typing import List from ltp import LTP from transformers.models.bert.tokenization_bert import BertTokenizer def a_ ( _lowercase ): # This defines a "chinese character" as anything in the CJK Unicode block: # https://en.wikipedia.org/wiki/CJK_Unified_Ideographs_(Unicode_block) # # Note that the CJK Unicode block is NOT all Japanese and Korean characters, # despite its name. The modern Korean Hangul alphabet is a different block, # as is Japanese Hiragana and Katakana. Those alphabets are used to write # space-separated words, so they are not treated specially and handled # like the all of the other languages. if ( (cp >= 0X4E00 and cp <= 0X9FFF) or (cp >= 0X3400 and cp <= 0X4DBF) # or (cp >= 0X20000 and cp <= 0X2A6DF) # or (cp >= 0X2A700 and cp <= 0X2B73F) # or (cp >= 0X2B740 and cp <= 0X2B81F) # or (cp >= 0X2B820 and cp <= 0X2CEAF) # or (cp >= 0XF900 and cp <= 0XFAFF) or (cp >= 0X2F800 and cp <= 0X2FA1F) # ): # return True return False def a_ ( _lowercase ): # word like '180' or '身高' or '神' for char in word: _UpperCamelCase : Dict = ord(_lowercase ) if not _is_chinese_char(_lowercase ): return 0 return 1 def a_ ( _lowercase ): _UpperCamelCase : List[str] = set() for token in tokens: _UpperCamelCase : int = len(_lowercase ) > 1 and is_chinese(_lowercase ) if chinese_word: word_set.add(_lowercase ) _UpperCamelCase : Optional[int] = list(_lowercase ) return word_list def a_ ( _lowercase , _lowercase ): if not chinese_word_set: return bert_tokens _UpperCamelCase : Tuple = max([len(_lowercase ) for w in chinese_word_set] ) _UpperCamelCase : int = bert_tokens _UpperCamelCase , _UpperCamelCase : Union[str, Any] = 0, len(_lowercase ) while start < end: _UpperCamelCase : Union[str, Any] = True if is_chinese(bert_word[start] ): _UpperCamelCase : List[Any] = min(end - start , _lowercase ) for i in range(_lowercase , 1 , -1 ): _UpperCamelCase : str = ''''''.join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1 , start + i ): _UpperCamelCase : int = '''##''' + bert_word[j] _UpperCamelCase : int = start + i _UpperCamelCase : Union[str, Any] = False break if single_word: start += 1 return bert_word def a_ ( _lowercase , _lowercase , _lowercase ): _UpperCamelCase : List[Any] = [] for i in range(0 , len(_lowercase ) , 100 ): _UpperCamelCase : Optional[int] = ltp_tokenizer.pipeline(lines[i : i + 100] , tasks=['''cws'''] ).cws _UpperCamelCase : Optional[int] = [get_chinese_word(_lowercase ) for r in res] ltp_res.extend(_lowercase ) assert len(_lowercase ) == len(_lowercase ) _UpperCamelCase : Dict = [] for i in range(0 , len(_lowercase ) , 100 ): _UpperCamelCase : Optional[int] = bert_tokenizer(lines[i : i + 100] , add_special_tokens=_lowercase , truncation=_lowercase , max_length=512 ) bert_res.extend(res['''input_ids'''] ) assert len(_lowercase ) == len(_lowercase ) _UpperCamelCase : Optional[Any] = [] for input_ids, chinese_word in zip(_lowercase , _lowercase ): _UpperCamelCase : str = [] for id in input_ids: _UpperCamelCase : Dict = bert_tokenizer._convert_id_to_token(_lowercase ) input_tokens.append(_lowercase ) _UpperCamelCase : str = add_sub_symbol(_lowercase , _lowercase ) _UpperCamelCase : List[str] = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(_lowercase ): if token[:2] == "##": _UpperCamelCase : int = token[2:] # save chinese tokens' pos if len(_lowercase ) == 1 and _is_chinese_char(ord(_lowercase ) ): ref_id.append(_lowercase ) ref_ids.append(_lowercase ) assert len(_lowercase ) == len(_lowercase ) return ref_ids def a_ ( _lowercase ): # For Chinese (Ro)Bert, the best result is from : RoBERTa-wwm-ext (https://github.com/ymcui/Chinese-BERT-wwm) # If we want to fine-tune these model, we have to use same tokenizer : LTP (https://github.com/HIT-SCIR/ltp) with open(args.file_name , '''r''' , encoding='''utf-8''' ) as f: _UpperCamelCase : Union[str, Any] = f.readlines() _UpperCamelCase : Tuple = [line.strip() for line in data if len(_lowercase ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' _UpperCamelCase : List[Any] = LTP(args.ltp ) # faster in GPU device _UpperCamelCase : int = BertTokenizer.from_pretrained(args.bert ) _UpperCamelCase : List[str] = prepare_ref(_lowercase , _lowercase , _lowercase ) with open(args.save_path , '''w''' , encoding='''utf-8''' ) as f: _UpperCamelCase : List[Any] = [json.dumps(_lowercase ) + '''\n''' for ref in ref_ids] f.writelines(_lowercase ) if __name__ == "__main__": UpperCamelCase_ =argparse.ArgumentParser(description="""prepare_chinese_ref""") parser.add_argument( """--file_name""", required=False, type=str, default="""./resources/chinese-demo.txt""", help="""file need process, same as training data in lm""", ) parser.add_argument( """--ltp""", required=False, type=str, default="""./resources/ltp""", help="""resources for LTP tokenizer, usually a path""", ) parser.add_argument( """--bert""", required=False, type=str, default="""./resources/robert""", help="""resources for Bert tokenizer""", ) parser.add_argument( """--save_path""", required=False, type=str, default="""./resources/ref.txt""", help="""path to save res""", ) UpperCamelCase_ =parser.parse_args() main(args)
128
0
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokenizer else: _lowerCamelCase =None _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ="""▁""" _lowerCamelCase ={"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} _lowerCamelCase ={ """vocab_file""": {"""google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model"""}, """tokenizer_file""": { """google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/tokenizer.json""" }, } _lowerCamelCase ={ """google/pegasus-xsum""": 5_1_2, } class A__ ( _lowerCAmelCase): _UpperCAmelCase : Tuple = VOCAB_FILES_NAMES _UpperCAmelCase : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase : Optional[Any] = PegasusTokenizer _UpperCAmelCase : Optional[int] = ["""input_ids""", """attention_mask"""] def __init__( self , __magic_name__=None , __magic_name__=None , __magic_name__="<pad>" , __magic_name__="</s>" , __magic_name__="<unk>" , __magic_name__="<mask_2>" , __magic_name__="<mask_1>" , __magic_name__=None , __magic_name__=1_0_3 , **__magic_name__ , ): lowerCamelCase : Dict = offset if additional_special_tokens is not None: if not isinstance(__magic_name__ , __magic_name__ ): raise TypeError( F'''additional_special_tokens should be of type {type(__magic_name__ )}, but is''' F''' {type(__magic_name__ )}''' ) lowerCamelCase : str = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ F'''<unk_{i}>''' for i in range(len(__magic_name__ ) , self.offset - 1 ) ] if len(set(__magic_name__ ) ) != len(__magic_name__ ): raise ValueError( """Please make sure that the provided additional_special_tokens do not contain an incorrectly""" F''' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.''' ) lowerCamelCase : Tuple = additional_special_tokens_extended else: lowerCamelCase : Dict = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [F'''<unk_{i}>''' for i in range(2 , self.offset )] super().__init__( __magic_name__ , tokenizer_file=__magic_name__ , pad_token=__magic_name__ , eos_token=__magic_name__ , unk_token=__magic_name__ , mask_token=__magic_name__ , mask_token_sent=__magic_name__ , offset=__magic_name__ , additional_special_tokens=__magic_name__ , **__magic_name__ , ) lowerCamelCase : Optional[int] = vocab_file lowerCamelCase : str = False if not self.vocab_file else True def UpperCamelCase__ ( self , __magic_name__ ): lowerCamelCase : Optional[Any] = set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special if all_special_ids != set(range(len(self.additional_special_tokens ) + 3 ) ): raise ValueError( """There should be 3 special tokens: mask_token, pad_token, and eos_token +""" F''' {len(self.additional_special_tokens )} additional_special_tokens, but got {all_special_ids}''' ) return [1 if x in all_special_ids else 0 for x in seq] def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = None , __magic_name__ = False ): if already_has_special_tokens: return self._special_token_mask(__magic_name__ ) elif token_ids_a is None: return self._special_token_mask(__magic_name__ ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def UpperCamelCase__ ( self , __magic_name__ , __magic_name__=None ): if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = None ): if not self.can_save_slow_tokenizer: raise ValueError( """Your fast tokenizer does not have the necessary information to save the vocabulary for a slow """ """tokenizer.""" ) if not os.path.isdir(__magic_name__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return lowerCamelCase : Optional[int] = os.path.join( __magic_name__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__magic_name__ ): copyfile(self.vocab_file , __magic_name__ ) return (out_vocab_file,)
287
"""simple docstring""" from __future__ import annotations from collections import Counter from random import random class lowerCamelCase : '''simple docstring''' def __init__( self: Tuple ) -> Optional[Any]: snake_case_ :Optional[int] = {} def lowerCAmelCase_ ( self: Dict , snake_case: str ) -> None: snake_case_ :str = {} def lowerCAmelCase_ ( self: Optional[int] , snake_case: str , snake_case: str , snake_case: float ) -> None: if nodea not in self.connections: self.add_node(snake_case ) if nodea not in self.connections: self.add_node(snake_case ) snake_case_ :Dict = probability def lowerCAmelCase_ ( self: List[Any] ) -> list[str]: return list(self.connections ) def lowerCAmelCase_ ( self: Any , snake_case: str ) -> str: snake_case_ :Optional[Any] = 0 snake_case_ :List[str] = random() for dest in self.connections[node]: current_probability += self.connections[node][dest] if current_probability > random_value: return dest return "" def A_ ( _lowercase, _lowercase, _lowercase ): '''simple docstring''' snake_case_ :List[str] = MarkovChainGraphUndirectedUnweighted() for nodea, nodea, probability in transitions: graph.add_transition_probability(_lowercase, _lowercase, _lowercase ) snake_case_ :int = Counter(graph.get_nodes() ) snake_case_ :Optional[Any] = start for _ in range(_lowercase ): snake_case_ :Tuple = graph.transition(_lowercase ) visited[node] += 1 return visited if __name__ == "__main__": import doctest doctest.testmod()
66
0
import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import DetrConfig, MaskFormerConfig, SwinConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskFormerForInstanceSegmentation, MaskFormerModel if is_vision_available(): from transformers import MaskFormerImageProcessor if is_vision_available(): from PIL import Image class __SCREAMING_SNAKE_CASE : def __init__( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=2 , SCREAMING_SNAKE_CASE__=True , SCREAMING_SNAKE_CASE__=False , SCREAMING_SNAKE_CASE__=10 , SCREAMING_SNAKE_CASE__=3 , SCREAMING_SNAKE_CASE__=32 * 4 , SCREAMING_SNAKE_CASE__=32 * 6 , SCREAMING_SNAKE_CASE__=4 , SCREAMING_SNAKE_CASE__=32 , ): lowercase : Dict = parent lowercase : Dict = batch_size lowercase : Union[str, Any] = is_training lowercase : Optional[Any] = use_auxiliary_loss lowercase : Optional[Any] = num_queries lowercase : Any = num_channels lowercase : List[Any] = min_size lowercase : int = max_size lowercase : str = num_labels lowercase : Union[str, Any] = mask_feature_size def __lowerCamelCase ( self ): lowercase : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( _SCREAMING_SNAKE_CASE ) lowercase : str = torch.ones([self.batch_size, self.min_size, self.max_size] , device=_SCREAMING_SNAKE_CASE ) lowercase : Any = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=_SCREAMING_SNAKE_CASE ) > 0.5 ).float() lowercase : Optional[int] = (torch.rand((self.batch_size, self.num_labels) , device=_SCREAMING_SNAKE_CASE ) > 0.5).long() lowercase : List[str] = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def __lowerCamelCase ( self ): return MaskFormerConfig.from_backbone_and_decoder_configs( backbone_config=SwinConfig( depths=[1, 1, 1, 1] , ) , decoder_config=DetrConfig( decoder_ffn_dim=128 , num_queries=self.num_queries , decoder_attention_heads=2 , d_model=self.mask_feature_size , ) , mask_feature_size=self.mask_feature_size , fpn_feature_size=self.mask_feature_size , num_channels=self.num_channels , num_labels=self.num_labels , ) def __lowerCamelCase ( self ): lowercase , lowercase , lowercase , lowercase , lowercase : Optional[Any] = self.prepare_config_and_inputs() lowercase : Optional[Any] = {'''pixel_values''': pixel_values, '''pixel_mask''': pixel_mask} return config, inputs_dict def __lowerCamelCase ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): lowercase : Dict = output.encoder_hidden_states lowercase : Union[str, Any] = output.pixel_decoder_hidden_states lowercase : List[Any] = output.transformer_decoder_hidden_states self.parent.assertTrue(len(_SCREAMING_SNAKE_CASE ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_SCREAMING_SNAKE_CASE ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_SCREAMING_SNAKE_CASE ) , config.decoder_config.decoder_layers ) def __lowerCamelCase ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False ): with torch.no_grad(): lowercase : List[str] = MaskFormerModel(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() lowercase : List[Any] = model(pixel_values=_SCREAMING_SNAKE_CASE , pixel_mask=_SCREAMING_SNAKE_CASE ) lowercase : Dict = model(_SCREAMING_SNAKE_CASE , output_hidden_states=_SCREAMING_SNAKE_CASE ) # the correct shape of output.transformer_decoder_hidden_states ensure the correcteness of the # encoder and pixel decoder self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.mask_feature_size) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def __lowerCamelCase ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): lowercase : str = MaskFormerForInstanceSegmentation(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() def comm_check_on_output(SCREAMING_SNAKE_CASE__ ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): lowercase : int = model(pixel_values=_SCREAMING_SNAKE_CASE , pixel_mask=_SCREAMING_SNAKE_CASE ) lowercase : Tuple = model(_SCREAMING_SNAKE_CASE ) comm_check_on_output(_SCREAMING_SNAKE_CASE ) lowercase : Any = model( pixel_values=_SCREAMING_SNAKE_CASE , pixel_mask=_SCREAMING_SNAKE_CASE , mask_labels=_SCREAMING_SNAKE_CASE , class_labels=_SCREAMING_SNAKE_CASE ) comm_check_on_output(_SCREAMING_SNAKE_CASE ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class __SCREAMING_SNAKE_CASE ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): A : Optional[int] = (MaskFormerModel, MaskFormerForInstanceSegmentation) if is_torch_available() else () A : Union[str, Any] = ( {'feature-extraction': MaskFormerModel, 'image-segmentation': MaskFormerForInstanceSegmentation} if is_torch_available() else {} ) A : List[Any] = False A : int = False A : Any = False A : Optional[int] = False def __lowerCamelCase ( self ): lowercase : str = MaskFormerModelTester(self ) lowercase : Dict = ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE , has_text_modality=_SCREAMING_SNAKE_CASE ) def __lowerCamelCase ( self ): self.config_tester.run_common_tests() def __lowerCamelCase ( self ): lowercase , lowercase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , output_hidden_states=_SCREAMING_SNAKE_CASE ) def __lowerCamelCase ( self ): lowercase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskformer_instance_segmentation_head_model(*_SCREAMING_SNAKE_CASE ) @unittest.skip(reason='''MaskFormer does not use inputs_embeds''' ) def __lowerCamelCase ( self ): pass @unittest.skip(reason='''MaskFormer does not have a get_input_embeddings method''' ) def __lowerCamelCase ( self ): pass @unittest.skip(reason='''MaskFormer is not a generative model''' ) def __lowerCamelCase ( self ): pass @unittest.skip(reason='''MaskFormer does not use token embeddings''' ) def __lowerCamelCase ( self ): pass @require_torch_multi_gpu @unittest.skip( reason='''MaskFormer has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`''' ) def __lowerCamelCase ( self ): pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def __lowerCamelCase ( self ): pass def __lowerCamelCase ( self ): lowercase , lowercase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase : Tuple = model_class(_SCREAMING_SNAKE_CASE ) lowercase : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase : List[Any] = [*signature.parameters.keys()] lowercase : Any = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , _SCREAMING_SNAKE_CASE ) @slow def __lowerCamelCase ( self ): for model_name in ["facebook/maskformer-swin-small-coco"]: lowercase : Union[str, Any] = MaskFormerModel.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) def __lowerCamelCase ( self ): lowercase : Union[str, Any] = (self.model_tester.min_size,) * 2 lowercase : Dict = { '''pixel_values''': torch.randn((2, 3, *size) , device=_SCREAMING_SNAKE_CASE ), '''mask_labels''': torch.randn((2, 10, *size) , device=_SCREAMING_SNAKE_CASE ), '''class_labels''': torch.zeros(2 , 10 , device=_SCREAMING_SNAKE_CASE ).long(), } lowercase : str = MaskFormerForInstanceSegmentation(MaskFormerConfig() ).to(_SCREAMING_SNAKE_CASE ) lowercase : Optional[int] = model(**_SCREAMING_SNAKE_CASE ) self.assertTrue(outputs.loss is not None ) def __lowerCamelCase ( self ): lowercase , lowercase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , output_hidden_states=_SCREAMING_SNAKE_CASE ) def __lowerCamelCase ( self ): lowercase , lowercase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase : Dict = model_class(_SCREAMING_SNAKE_CASE ).to(_SCREAMING_SNAKE_CASE ) lowercase : Union[str, Any] = model(**_SCREAMING_SNAKE_CASE , output_attentions=_SCREAMING_SNAKE_CASE ) self.assertTrue(outputs.attentions is not None ) def __lowerCamelCase ( self ): if not self.model_tester.is_training: return # only MaskFormerForInstanceSegmentation has the loss lowercase : List[str] = self.all_model_classes[1] lowercase , lowercase , lowercase , lowercase , lowercase : Tuple = self.model_tester.prepare_config_and_inputs() lowercase : Optional[Any] = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.train() lowercase : Union[str, Any] = model(_SCREAMING_SNAKE_CASE , mask_labels=_SCREAMING_SNAKE_CASE , class_labels=_SCREAMING_SNAKE_CASE ).loss loss.backward() def __lowerCamelCase ( self ): # only MaskFormerForInstanceSegmentation has the loss lowercase : List[Any] = self.all_model_classes[1] lowercase , lowercase , lowercase , lowercase , lowercase : str = self.model_tester.prepare_config_and_inputs() lowercase : str = True lowercase : Any = True lowercase : Dict = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.train() lowercase : str = model(_SCREAMING_SNAKE_CASE , mask_labels=_SCREAMING_SNAKE_CASE , class_labels=_SCREAMING_SNAKE_CASE ) lowercase : int = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() lowercase : List[Any] = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() # we requires_grad=True in inputs_embeds (line 2152), the original implementation don't lowercase : List[Any] = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() lowercase : int = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) __a = 1e-4 def __lowercase ( ) ->List[Any]: """simple docstring""" lowercase : Any = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_vision @slow class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @cached_property def __lowerCamelCase ( self ): return ( MaskFormerImageProcessor.from_pretrained('''facebook/maskformer-swin-small-coco''' ) if is_vision_available() else None ) def __lowerCamelCase ( self ): lowercase : Any = MaskFormerModel.from_pretrained('''facebook/maskformer-swin-small-coco''' ).to(_SCREAMING_SNAKE_CASE ) lowercase : Dict = self.default_image_processor lowercase : Optional[int] = prepare_img() lowercase : Any = image_processor(_SCREAMING_SNAKE_CASE , return_tensors='''pt''' ).to(_SCREAMING_SNAKE_CASE ) lowercase : str = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_SCREAMING_SNAKE_CASE , (1, 3, 800, 1088) ) with torch.no_grad(): lowercase : Union[str, Any] = model(**_SCREAMING_SNAKE_CASE ) lowercase : int = torch.tensor( [[-0.0482, 0.9228, 0.4951], [-0.2547, 0.8017, 0.8527], [-0.0069, 0.3385, -0.0089]] ).to(_SCREAMING_SNAKE_CASE ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , _SCREAMING_SNAKE_CASE , atol=_SCREAMING_SNAKE_CASE ) ) lowercase : List[Any] = torch.tensor( [[-0.8422, -0.8434, -0.9718], [-1.0144, -0.5565, -0.4195], [-1.0038, -0.4484, -0.1961]] ).to(_SCREAMING_SNAKE_CASE ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , _SCREAMING_SNAKE_CASE , atol=_SCREAMING_SNAKE_CASE ) ) lowercase : Any = torch.tensor( [[0.2852, -0.0159, 0.9735], [0.6254, 0.1858, 0.8529], [-0.0680, -0.4116, 1.8413]] ).to(_SCREAMING_SNAKE_CASE ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , _SCREAMING_SNAKE_CASE , atol=_SCREAMING_SNAKE_CASE ) ) def __lowerCamelCase ( self ): lowercase : Optional[int] = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-swin-small-coco''' ) .to(_SCREAMING_SNAKE_CASE ) .eval() ) lowercase : List[str] = self.default_image_processor lowercase : Union[str, Any] = prepare_img() lowercase : Optional[Any] = image_processor(_SCREAMING_SNAKE_CASE , return_tensors='''pt''' ).to(_SCREAMING_SNAKE_CASE ) lowercase : List[Any] = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_SCREAMING_SNAKE_CASE , (1, 3, 800, 1088) ) with torch.no_grad(): lowercase : Dict = model(**_SCREAMING_SNAKE_CASE ) # masks_queries_logits lowercase : Dict = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) lowercase : int = [ [-1.3737124, -1.7724937, -1.9364233], [-1.5977281, -1.9867939, -2.1523695], [-1.5795398, -1.9269832, -2.093942], ] lowercase : Dict = torch.tensor(_SCREAMING_SNAKE_CASE ).to(_SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , _SCREAMING_SNAKE_CASE , atol=_SCREAMING_SNAKE_CASE ) ) # class_queries_logits lowercase : Union[str, Any] = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) lowercase : Optional[Any] = torch.tensor( [ [1.6512E00, -5.2572E00, -3.3519E00], [3.6169E-02, -5.9025E00, -2.9313E00], [1.0766E-04, -7.7630E00, -5.1263E00], ] ).to(_SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , _SCREAMING_SNAKE_CASE , atol=_SCREAMING_SNAKE_CASE ) ) def __lowerCamelCase ( self ): lowercase : List[str] = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-resnet101-coco-stuff''' ) .to(_SCREAMING_SNAKE_CASE ) .eval() ) lowercase : Optional[Any] = self.default_image_processor lowercase : Optional[int] = prepare_img() lowercase : Any = image_processor(_SCREAMING_SNAKE_CASE , return_tensors='''pt''' ).to(_SCREAMING_SNAKE_CASE ) lowercase : Optional[int] = inputs['''pixel_values'''].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_SCREAMING_SNAKE_CASE , (1, 3, 800, 1088) ) with torch.no_grad(): lowercase : Union[str, Any] = model(**_SCREAMING_SNAKE_CASE ) # masks_queries_logits lowercase : Optional[Any] = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) lowercase : List[Any] = [[-0.9046, -2.6366, -4.6062], [-3.4179, -5.7890, -8.8057], [-4.9179, -7.6560, -10.7711]] lowercase : List[str] = torch.tensor(_SCREAMING_SNAKE_CASE ).to(_SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , _SCREAMING_SNAKE_CASE , atol=_SCREAMING_SNAKE_CASE ) ) # class_queries_logits lowercase : Union[str, Any] = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) lowercase : List[str] = torch.tensor( [[4.7188, -3.2585, -2.8857], [6.6871, -2.9181, -1.2487], [7.2449, -2.2764, -2.1874]] ).to(_SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , _SCREAMING_SNAKE_CASE , atol=_SCREAMING_SNAKE_CASE ) ) def __lowerCamelCase ( self ): lowercase : List[str] = ( MaskFormerForInstanceSegmentation.from_pretrained('''facebook/maskformer-swin-small-coco''' ) .to(_SCREAMING_SNAKE_CASE ) .eval() ) lowercase : Any = self.default_image_processor lowercase : Union[str, Any] = image_processor( [np.zeros((3, 800, 1333) ), np.zeros((3, 800, 1333) )] , segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] , return_tensors='''pt''' , ) lowercase : Optional[Any] = inputs['''pixel_values'''].to(_SCREAMING_SNAKE_CASE ) lowercase : List[str] = [el.to(_SCREAMING_SNAKE_CASE ) for el in inputs['''mask_labels''']] lowercase : Tuple = [el.to(_SCREAMING_SNAKE_CASE ) for el in inputs['''class_labels''']] with torch.no_grad(): lowercase : str = model(**_SCREAMING_SNAKE_CASE ) self.assertTrue(outputs.loss is not None )
368
# Algorithm for the pigeonhole sorting def __lowercase ( _UpperCamelCase ) ->List[Any]: """simple docstring""" lowercase : List[Any] = min(_UpperCamelCase ) # min() finds the minimum value lowercase : Union[str, Any] = max(_UpperCamelCase ) # max() finds the maximum value lowercase : Tuple = max_val - min_val + 1 # size is difference of max and min values plus one # list of pigeonholes of size equal to the variable size lowercase : List[Any] = [0] * size # Populate the pigeonholes. for x in a: assert isinstance(_UpperCamelCase, _UpperCamelCase ), "integers only please" holes[x - min_val] += 1 # Putting the elements back into the array in an order. lowercase : Tuple = 0 for count in range(_UpperCamelCase ): while holes[count] > 0: holes[count] -= 1 lowercase : str = count + min_val i += 1 def __lowercase ( ) ->List[str]: """simple docstring""" lowercase : Union[str, Any] = [8, 3, 2, 7, 4, 6, 8] pigeonhole_sort(_UpperCamelCase ) print('''Sorted order is:''', ''' '''.join(_UpperCamelCase ) ) if __name__ == "__main__": main()
173
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __snake_case : Any = { """configuration_clipseg""": [ """CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP""", """CLIPSegConfig""", """CLIPSegTextConfig""", """CLIPSegVisionConfig""", ], """processing_clipseg""": ["""CLIPSegProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : int = [ """CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST""", """CLIPSegModel""", """CLIPSegPreTrainedModel""", """CLIPSegTextModel""", """CLIPSegVisionModel""", """CLIPSegForImageSegmentation""", ] if TYPE_CHECKING: from .configuration_clipseg import ( CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP, CLIPSegConfig, CLIPSegTextConfig, CLIPSegVisionConfig, ) from .processing_clipseg import CLIPSegProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_clipseg import ( CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST, CLIPSegForImageSegmentation, CLIPSegModel, CLIPSegPreTrainedModel, CLIPSegTextModel, CLIPSegVisionModel, ) else: import sys __snake_case : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
248
from dataclasses import dataclass from typing import Optional, Tuple import torch from torch import nn from transformers import RobertaPreTrainedModel, XLMRobertaConfig, XLMRobertaModel from transformers.utils import ModelOutput @dataclass class A__(a_ ): """simple docstring""" _A : Optional[torch.FloatTensor] = None _A : torch.FloatTensor = None _A : Optional[Tuple[torch.FloatTensor]] = None _A : Optional[Tuple[torch.FloatTensor]] = None class A__(a_ ): """simple docstring""" def __init__( self , _lowercase=1 , _lowercase=0 , _lowercase=2 , _lowercase=512 , _lowercase="cls" , _lowercase=False , _lowercase=True , **_lowercase , ) -> Union[str, Any]: super().__init__(pad_token_id=_lowercase , bos_token_id=_lowercase , eos_token_id=_lowercase , **_lowercase ) a_ : str = project_dim a_ : List[Any] = pooler_fn a_ : Union[str, Any] = learn_encoder a_ : List[str] = use_attention_mask class A__(a_ ): """simple docstring""" _A : Any = [r'''pooler''', r'''logit_scale'''] _A : List[str] = [r'''position_ids''', r'''predictions.decoder.bias'''] _A : List[str] = '''roberta''' _A : Union[str, Any] = RobertaSeriesConfig def __init__( self , _lowercase ) -> Optional[Any]: super().__init__(_lowercase ) a_ : Optional[int] = XLMRobertaModel(_lowercase ) a_ : Any = nn.Linear(config.hidden_size , config.project_dim ) a_ : Union[str, Any] = getattr(_lowercase , """has_pre_transformation""" , _lowercase ) if self.has_pre_transformation: a_ : int = nn.Linear(config.hidden_size , config.project_dim ) a_ : Union[str, Any] = nn.LayerNorm(config.hidden_size , eps=config.layer_norm_eps ) self.post_init() def UpperCamelCase__ ( self , _lowercase = None , _lowercase = None , _lowercase = None , _lowercase = None , _lowercase = None , _lowercase = None , _lowercase = None , _lowercase = None , _lowercase = None , _lowercase = None , _lowercase = None , ) -> Any: a_ : str = return_dict if return_dict is not None else self.config.use_return_dict a_ : Any = self.base_model( input_ids=_lowercase , attention_mask=_lowercase , token_type_ids=_lowercase , position_ids=_lowercase , head_mask=_lowercase , inputs_embeds=_lowercase , encoder_hidden_states=_lowercase , encoder_attention_mask=_lowercase , output_attentions=_lowercase , output_hidden_states=True if self.has_pre_transformation else output_hidden_states , return_dict=_lowercase , ) if self.has_pre_transformation: a_ : str = outputs["""hidden_states"""][-2] a_ : Tuple = self.pre_LN(_lowercase ) a_ : List[str] = self.transformation_pre(_lowercase ) return TransformationModelOutput( projection_state=_lowercase , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , ) else: a_ : Union[str, Any] = self.transformation(outputs.last_hidden_state ) return TransformationModelOutput( projection_state=_lowercase , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
248
1
'''simple docstring''' def lowerCAmelCase__ ( lowerCamelCase : list ): _A : int = False while is_sorted is False: # Until all the indices are traversed keep looping _A : str = True for i in range(0 ,len(lowerCamelCase ) - 1 ,2 ): # iterating over all even indices if input_list[i] > input_list[i + 1]: _A , _A : Any = input_list[i + 1], input_list[i] # swapping if elements not in order _A : str = False for i in range(1 ,len(lowerCamelCase ) - 1 ,2 ): # iterating over all odd indices if input_list[i] > input_list[i + 1]: _A , _A : Tuple = input_list[i + 1], input_list[i] # swapping if elements not in order _A : int = False return input_list if __name__ == "__main__": print('''Enter list to be sorted''') A : List[str] = [int(x) for x in input().split()] # inputing elements of the list in one line A : Optional[Any] = odd_even_sort(input_list) print('''The sorted list is''') print(sorted_list)
227
'''simple docstring''' import operator as op A : List[str] = '''scaler.pt''' A : Optional[int] = '''pytorch_model''' A : Union[str, Any] = '''random_states''' A : List[Any] = '''optimizer''' A : Optional[Any] = '''scheduler''' A : Optional[int] = '''pytorch_model.bin''' A : List[str] = '''pytorch_model.bin.index.json''' A : List[Any] = '''model.safetensors''' A : int = '''model.safetensors.index.json''' A : List[str] = '''1.10.2''' A : str = '''py38''' A : int = '''4.17.0''' A : Tuple = ['''ml.p3.16xlarge''', '''ml.p3dn.24xlarge''', '''ml.p4dn.24xlarge'''] A : Union[str, Any] = ['''FULL_SHARD''', '''SHARD_GRAD_OP''', '''NO_SHARD''', '''HYBRID_SHARD''', '''HYBRID_SHARD_ZERO2'''] A : Optional[Any] = ['''TRANSFORMER_BASED_WRAP''', '''SIZE_BASED_WRAP''', '''NO_WRAP'''] A : Dict = ['''BACKWARD_PRE''', '''BACKWARD_POST''', '''NO_PREFETCH'''] A : Any = ['''FULL_STATE_DICT''', '''LOCAL_STATE_DICT''', '''SHARDED_STATE_DICT'''] A : Tuple = '''2.0.1''' A : List[str] = ['''pdsh''', '''standard''', '''openmpi''', '''mvapich'''] A : Union[str, Any] = ['''default''', '''reduce-overhead''', '''max-autotune'''] A : List[Any] = {'''>''': op.gt, '''>=''': op.ge, '''==''': op.eq, '''!=''': op.ne, '''<=''': op.le, '''<''': op.lt} # These are the args for `torch.distributed.launch` for pytorch < 1.9 A : Tuple = [ '''nnodes''', '''nproc_per_node''', '''rdzv_backend''', '''rdzv_endpoint''', '''rdzv_id''', '''rdzv_conf''', '''standalone''', '''max_restarts''', '''monitor_interval''', '''start_method''', '''role''', '''module''', '''m''', '''no_python''', '''run_path''', '''log_dir''', '''r''', '''redirects''', '''t''', '''tee''', '''node_rank''', '''master_addr''', '''master_port''', ] A : List[str] = ['''DEEPSPEED''', '''MULTI_GPU''', '''FSDP''', '''MEGATRON_LM'''] A : Any = ['''DEEPSPEED''', '''MULTI_XPU''', '''FSDP''']
227
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __A = {"configuration_ibert": ["IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "IBertConfig", "IBertOnnxConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "IBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "IBertForMaskedLM", "IBertForMultipleChoice", "IBertForQuestionAnswering", "IBertForSequenceClassification", "IBertForTokenClassification", "IBertModel", "IBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys __A = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
10
import math class _UpperCamelCase : def __init__( self :Union[str, Any] , lowerCamelCase :Union[str, Any]=0 ) -> Tuple: # a graph with Node 0,1,...,N-1 UpperCAmelCase__ = n UpperCAmelCase__ = [ [math.inf for j in range(0 , lowerCamelCase )] for i in range(0 , lowerCamelCase ) ] # adjacency matrix for weight UpperCAmelCase__ = [ [math.inf for j in range(0 , lowerCamelCase )] for i in range(0 , lowerCamelCase ) ] # dp[i][j] stores minimum distance from i to j def UpperCAmelCase_ ( self :Tuple , lowerCamelCase :List[Any] , lowerCamelCase :Optional[Any] , lowerCamelCase :int ) -> List[Any]: UpperCAmelCase__ = w def UpperCAmelCase_ ( self :Optional[int] ) -> Optional[Any]: for k in range(0 , self.n ): for i in range(0 , self.n ): for j in range(0 , self.n ): UpperCAmelCase__ = min(self.dp[i][j] , self.dp[i][k] + self.dp[k][j] ) def UpperCAmelCase_ ( self :int , lowerCamelCase :List[Any] , lowerCamelCase :Dict ) -> List[str]: return self.dp[u][v] if __name__ == "__main__": _lowerCAmelCase : Optional[int] = Graph(5) graph.add_edge(0, 2, 9) graph.add_edge(0, 4, 1_0) graph.add_edge(1, 3, 5) graph.add_edge(2, 3, 7) graph.add_edge(3, 0, 1_0) graph.add_edge(3, 1, 2) graph.add_edge(3, 2, 1) graph.add_edge(3, 4, 6) graph.add_edge(4, 1, 3) graph.add_edge(4, 2, 4) graph.add_edge(4, 3, 9) graph.floyd_warshall() graph.show_min(1, 4) graph.show_min(0, 3)
169
0
import unittest import numpy as np from transformers import MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING, TF_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING from transformers.pipelines import AudioClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_torchaudio, slow, ) from .test_pipelines_common import ANY @is_pipeline_test class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' __UpperCamelCase : List[str] = MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING __UpperCamelCase : Any = TF_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING def __magic_name__ ( self : str , lowerCAmelCase_ : int , lowerCAmelCase_ : Dict , lowerCAmelCase_ : Dict ): """simple docstring""" _A: Union[str, Any] = AudioClassificationPipeline(model=lowerCAmelCase_ , feature_extractor=lowerCAmelCase_ ) # test with a raw waveform _A: Optional[Any] = np.zeros((3_4_0_0_0,) ) _A: int = np.zeros((1_4_0_0_0,) ) return audio_classifier, [audioa, audio] def __magic_name__ ( self : int , lowerCAmelCase_ : Dict , lowerCAmelCase_ : Any ): """simple docstring""" _A , _A: Any = examples _A: List[str] = audio_classifier(lowerCAmelCase_ ) # by default a model is initialized with num_labels=2 self.assertEqual( lowerCAmelCase_ , [ {'''score''': ANY(lowerCAmelCase_ ), '''label''': ANY(lowerCAmelCase_ )}, {'''score''': ANY(lowerCAmelCase_ ), '''label''': ANY(lowerCAmelCase_ )}, ] , ) _A: Optional[Any] = audio_classifier(lowerCAmelCase_ , top_k=1 ) self.assertEqual( lowerCAmelCase_ , [ {'''score''': ANY(lowerCAmelCase_ ), '''label''': ANY(lowerCAmelCase_ )}, ] , ) self.run_torchaudio(lowerCAmelCase_ ) @require_torchaudio def __magic_name__ ( self : Any , lowerCAmelCase_ : Optional[int] ): """simple docstring""" import datasets # test with a local file _A: List[str] = datasets.load_dataset('''hf-internal-testing/librispeech_asr_dummy''' , '''clean''' , split='''validation''' ) _A: Any = dataset[0]['''audio''']['''array'''] _A: Optional[int] = audio_classifier(lowerCAmelCase_ ) self.assertEqual( lowerCAmelCase_ , [ {'''score''': ANY(lowerCAmelCase_ ), '''label''': ANY(lowerCAmelCase_ )}, {'''score''': ANY(lowerCAmelCase_ ), '''label''': ANY(lowerCAmelCase_ )}, ] , ) @require_torch def __magic_name__ ( self : List[Any] ): """simple docstring""" _A: Optional[Any] = '''anton-l/wav2vec2-random-tiny-classifier''' _A: Dict = pipeline('''audio-classification''' , model=lowerCAmelCase_ ) _A: List[str] = np.ones((8_0_0_0,) ) _A: str = audio_classifier(lowerCAmelCase_ , top_k=4 ) _A: Tuple = [ {'''score''': 0.0842, '''label''': '''no'''}, {'''score''': 0.0838, '''label''': '''up'''}, {'''score''': 0.0837, '''label''': '''go'''}, {'''score''': 0.0834, '''label''': '''right'''}, ] _A: str = [ {'''score''': 0.0845, '''label''': '''stop'''}, {'''score''': 0.0844, '''label''': '''on'''}, {'''score''': 0.0841, '''label''': '''right'''}, {'''score''': 0.0834, '''label''': '''left'''}, ] self.assertIn(nested_simplify(lowerCAmelCase_ , decimals=4 ) , [EXPECTED_OUTPUT, EXPECTED_OUTPUT_PT_2] ) _A: Optional[int] = {'''array''': np.ones((8_0_0_0,) ), '''sampling_rate''': audio_classifier.feature_extractor.sampling_rate} _A: List[str] = audio_classifier(lowerCAmelCase_ , top_k=4 ) self.assertIn(nested_simplify(lowerCAmelCase_ , decimals=4 ) , [EXPECTED_OUTPUT, EXPECTED_OUTPUT_PT_2] ) @require_torch @slow def __magic_name__ ( self : str ): """simple docstring""" import datasets _A: Union[str, Any] = '''superb/wav2vec2-base-superb-ks''' _A: int = pipeline('''audio-classification''' , model=lowerCAmelCase_ ) _A: Union[str, Any] = datasets.load_dataset('''anton-l/superb_dummy''' , '''ks''' , split='''test''' ) _A: Optional[Any] = np.array(dataset[3]['''speech'''] , dtype=np.floataa ) _A: Tuple = audio_classifier(lowerCAmelCase_ , top_k=4 ) self.assertEqual( nested_simplify(lowerCAmelCase_ , decimals=3 ) , [ {'''score''': 0.981, '''label''': '''go'''}, {'''score''': 0.007, '''label''': '''up'''}, {'''score''': 0.006, '''label''': '''_unknown_'''}, {'''score''': 0.001, '''label''': '''down'''}, ] , ) @require_tf @unittest.skip('''Audio classification is not implemented for TF''' ) def __magic_name__ ( self : List[Any] ): """simple docstring""" pass
301
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase__ : Tuple = logging.get_logger(__name__) UpperCAmelCase__ : Union[str, Any] = { 'google/mobilenet_v1_1.0_224': 'https://huggingface.co/google/mobilenet_v1_1.0_224/resolve/main/config.json', 'google/mobilenet_v1_0.75_192': 'https://huggingface.co/google/mobilenet_v1_0.75_192/resolve/main/config.json', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 } class UpperCAmelCase ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' __UpperCamelCase : Optional[int] = '''mobilenet_v1''' def __init__( self : Optional[int] , lowerCAmelCase_ : Any=3 , lowerCAmelCase_ : str=2_2_4 , lowerCAmelCase_ : List[str]=1.0 , lowerCAmelCase_ : Any=8 , lowerCAmelCase_ : Tuple="relu6" , lowerCAmelCase_ : List[Any]=True , lowerCAmelCase_ : Optional[int]=0.999 , lowerCAmelCase_ : List[str]=0.02 , lowerCAmelCase_ : List[Any]=0.001 , **lowerCAmelCase_ : Optional[Any] , ): """simple docstring""" super().__init__(**lowerCAmelCase_ ) if depth_multiplier <= 0: raise ValueError('''depth_multiplier must be greater than zero.''' ) _A: Any = num_channels _A: Optional[int] = image_size _A: Optional[Any] = depth_multiplier _A: Tuple = min_depth _A: Any = hidden_act _A: Dict = tf_padding _A: List[Any] = classifier_dropout_prob _A: Tuple = initializer_range _A: Tuple = layer_norm_eps class UpperCAmelCase ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' __UpperCamelCase : Dict = version.parse('''1.11''' ) @property def __magic_name__ ( self : Union[str, Any] ): """simple docstring""" return OrderedDict([('''pixel_values''', {0: '''batch'''})] ) @property def __magic_name__ ( self : Optional[Any] ): """simple docstring""" if self.task == "image-classification": return OrderedDict([('''logits''', {0: '''batch'''})] ) else: return OrderedDict([('''last_hidden_state''', {0: '''batch'''}), ('''pooler_output''', {0: '''batch'''})] ) @property def __magic_name__ ( self : Dict ): """simple docstring""" return 1e-4
301
1
'''simple docstring''' import os import string import sys __lowerCAmelCase = 1 << 8 __lowerCAmelCase = { """tab""": ord("""\t"""), """newline""": ord("""\r"""), """esc""": 2_7, """up""": 6_5 + ARROW_KEY_FLAG, """down""": 6_6 + ARROW_KEY_FLAG, """right""": 6_7 + ARROW_KEY_FLAG, """left""": 6_8 + ARROW_KEY_FLAG, """mod_int""": 9_1, """undefined""": sys.maxsize, """interrupt""": 3, """insert""": 5_0, """delete""": 5_1, """pg_up""": 5_3, """pg_down""": 5_4, } __lowerCAmelCase = KEYMAP["""up"""] __lowerCAmelCase = KEYMAP["""left"""] if sys.platform == "win32": __lowerCAmelCase = [] __lowerCAmelCase = { B"""\xe0H""": KEYMAP["""up"""] - ARROW_KEY_FLAG, B"""\x00H""": KEYMAP["""up"""] - ARROW_KEY_FLAG, B"""\xe0P""": KEYMAP["""down"""] - ARROW_KEY_FLAG, B"""\x00P""": KEYMAP["""down"""] - ARROW_KEY_FLAG, B"""\xe0M""": KEYMAP["""right"""] - ARROW_KEY_FLAG, B"""\x00M""": KEYMAP["""right"""] - ARROW_KEY_FLAG, B"""\xe0K""": KEYMAP["""left"""] - ARROW_KEY_FLAG, B"""\x00K""": KEYMAP["""left"""] - ARROW_KEY_FLAG, } for i in range(1_0): __lowerCAmelCase = ord(str(i)) def UpperCAmelCase_ (): """simple docstring""" if os.name == "nt": import msvcrt _a : List[Any] = 'mbcs' # Flush the keyboard buffer while msvcrt.kbhit(): msvcrt.getch() if len(__a ) == 0: # Read the keystroke _a : int = msvcrt.getch() # If it is a prefix char, get second part if ch in (b"\x00", b"\xe0"): _a : List[Any] = ch + msvcrt.getch() # Translate actual Win chars to bullet char types try: _a : int = chr(WIN_KEYMAP[cha] ) WIN_CH_BUFFER.append(chr(KEYMAP['mod_int'] ) ) WIN_CH_BUFFER.append(__a ) if ord(__a ) in ( KEYMAP["insert"] - 1 << 9, KEYMAP["delete"] - 1 << 9, KEYMAP["pg_up"] - 1 << 9, KEYMAP["pg_down"] - 1 << 9, ): WIN_CH_BUFFER.append(chr(1_2_6 ) ) _a : Any = chr(KEYMAP['esc'] ) except KeyError: _a : int = cha[1] else: _a : Any = ch.decode(__a ) else: _a : int = WIN_CH_BUFFER.pop(0 ) elif os.name == "posix": import termios import tty _a : Optional[Any] = sys.stdin.fileno() _a : List[Any] = termios.tcgetattr(__a ) try: tty.setraw(__a ) _a : List[Any] = sys.stdin.read(1 ) finally: termios.tcsetattr(__a , termios.TCSADRAIN , __a ) return ch def UpperCAmelCase_ (): """simple docstring""" _a : List[str] = get_raw_chars() if ord(__a ) in [KEYMAP["interrupt"], KEYMAP["newline"]]: return char elif ord(__a ) == KEYMAP["esc"]: _a : int = get_raw_chars() if ord(__a ) == KEYMAP["mod_int"]: _a : str = get_raw_chars() if ord(__a ) >= KEYMAP["arrow_begin"] - ARROW_KEY_FLAG and ord(__a ) <= KEYMAP["arrow_end"] - ARROW_KEY_FLAG: return chr(ord(__a ) + ARROW_KEY_FLAG ) else: return KEYMAP["undefined"] else: return get_raw_chars() else: if char in string.printable: return char else: return KEYMAP["undefined"]
271
'''simple docstring''' import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser __lowerCAmelCase = re.compile(r"""\s+""") def UpperCAmelCase_ (__a : Any ): """simple docstring""" return {"hash": hashlib.mda(re.sub(__a , '' , example['content'] ).encode('utf-8' ) ).hexdigest()} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : List[str] = [len(__a ) for line in example['content'].splitlines()] return {"line_mean": np.mean(__a ), "line_max": max(__a )} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Union[str, Any] = np.mean([c.isalnum() for c in example['content']] ) return {"alpha_frac": alpha_frac} def UpperCAmelCase_ (__a : Optional[int] , __a : Any ): """simple docstring""" if example["hash"] in uniques: uniques.remove(example['hash'] ) return True else: return False def UpperCAmelCase_ (__a : int , __a : Union[str, Any]=5 ): """simple docstring""" _a : Optional[int] = ['auto-generated', 'autogenerated', 'automatically generated'] _a : List[str] = example['content'].splitlines() for _, line in zip(range(__a ) , __a ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def UpperCAmelCase_ (__a : List[str] , __a : Dict=5 , __a : Tuple=0.05 ): """simple docstring""" _a : Optional[int] = ['unit tests', 'test file', 'configuration file'] _a : int = example['content'].splitlines() _a : int = 0 _a : Dict = 0 # first test for _, line in zip(range(__a ) , __a ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test _a : int = example['content'].count('\n' ) _a : int = int(coeff * nlines ) for line in lines: count_config += line.lower().count('config' ) count_test += line.lower().count('test' ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def UpperCAmelCase_ (__a : Optional[int] ): """simple docstring""" _a : List[str] = ['def ', 'class ', 'for ', 'while '] _a : str = example['content'].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def UpperCAmelCase_ (__a : int , __a : Any=4 ): """simple docstring""" _a : List[str] = example['content'].splitlines() _a : Dict = 0 for line in lines: counter += line.lower().count('=' ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Optional[Any] = tokenizer(example['content'] , truncation=__a )['input_ids'] _a : Optional[int] = len(example['content'] ) / len(__a ) return {"ratio": ratio} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Dict = {} results.update(get_hash(__a ) ) results.update(line_stats(__a ) ) results.update(alpha_stats(__a ) ) results.update(char_token_ratio(__a ) ) results.update(is_autogenerated(__a ) ) results.update(is_config_or_test(__a ) ) results.update(has_no_keywords(__a ) ) results.update(has_few_assignments(__a ) ) return results def UpperCAmelCase_ (__a : Any , __a : Any , __a : str ): """simple docstring""" if not check_uniques(__a , __a ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def UpperCAmelCase_ (__a : Union[str, Any] ): """simple docstring""" with open(__a , 'rb' ) as f_in: with gzip.open(str(__a ) + '.gz' , 'wb' , compresslevel=6 ) as f_out: shutil.copyfileobj(__a , __a ) os.unlink(__a ) # Settings __lowerCAmelCase = HfArgumentParser(PreprocessingArguments) __lowerCAmelCase = parser.parse_args() if args.num_workers is None: __lowerCAmelCase = multiprocessing.cpu_count() __lowerCAmelCase = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset __lowerCAmelCase = time.time() __lowerCAmelCase = load_dataset(args.dataset_name, split="""train""") print(f'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing __lowerCAmelCase = time.time() __lowerCAmelCase = ds.map(preprocess, num_proc=args.num_workers) print(f'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes __lowerCAmelCase = set(ds.unique("""hash""")) __lowerCAmelCase = len(uniques) / len(ds) print(f'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics __lowerCAmelCase = time.time() __lowerCAmelCase = ds.filter(filter, fn_kwargs={"""uniques""": uniques, """args""": args}) print(f'''Time to filter dataset: {time.time()-t_start:.2f}''') print(f'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: __lowerCAmelCase = time.time() __lowerCAmelCase , __lowerCAmelCase = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(f'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(f'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file __lowerCAmelCase = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / """duplicate_clusters.json""", """w""") as f: json.dump(duplicate_clusters, f) __lowerCAmelCase = output_dir / """data""" data_dir.mkdir(exist_ok=True) __lowerCAmelCase = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): __lowerCAmelCase = str(data_dir / f'''file-{file_number+1:012}.json''') __lowerCAmelCase = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(f'''Time to save dataset: {time.time()-t_start:.2f}''')
271
1
from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case : List[str] =logging.get_logger(__name__) __snake_case : Union[str, Any] ={ 'microsoft/markuplm-base': 'https://huggingface.co/microsoft/markuplm-base/resolve/main/config.json', 'microsoft/markuplm-large': 'https://huggingface.co/microsoft/markuplm-large/resolve/main/config.json', } class lowerCamelCase__ ( lowerCamelCase__): '''simple docstring''' snake_case_ ="""markuplm""" def __init__(self ,__lowerCamelCase=3_05_22 ,__lowerCamelCase=7_68 ,__lowerCamelCase=12 ,__lowerCamelCase=12 ,__lowerCamelCase=30_72 ,__lowerCamelCase="gelu" ,__lowerCamelCase=0.1 ,__lowerCamelCase=0.1 ,__lowerCamelCase=5_12 ,__lowerCamelCase=2 ,__lowerCamelCase=0.02 ,__lowerCamelCase=1e-12 ,__lowerCamelCase=0 ,__lowerCamelCase=0 ,__lowerCamelCase=2 ,__lowerCamelCase=2_56 ,__lowerCamelCase=10_24 ,__lowerCamelCase=2_16 ,__lowerCamelCase=10_01 ,__lowerCamelCase=32 ,__lowerCamelCase=50 ,__lowerCamelCase="absolute" ,__lowerCamelCase=True ,__lowerCamelCase=None ,**__lowerCamelCase ,) -> List[Any]: """simple docstring""" super().__init__( pad_token_id=__lowerCamelCase ,bos_token_id=__lowerCamelCase ,eos_token_id=__lowerCamelCase ,**__lowerCamelCase ,) lowerCAmelCase__ : int = vocab_size lowerCAmelCase__ : Optional[Any] = hidden_size lowerCAmelCase__ : Any = num_hidden_layers lowerCAmelCase__ : Optional[int] = num_attention_heads lowerCAmelCase__ : str = hidden_act lowerCAmelCase__ : Tuple = intermediate_size lowerCAmelCase__ : Any = hidden_dropout_prob lowerCAmelCase__ : Tuple = attention_probs_dropout_prob lowerCAmelCase__ : List[str] = max_position_embeddings lowerCAmelCase__ : Union[str, Any] = type_vocab_size lowerCAmelCase__ : List[str] = initializer_range lowerCAmelCase__ : str = layer_norm_eps lowerCAmelCase__ : Union[str, Any] = position_embedding_type lowerCAmelCase__ : List[Any] = use_cache lowerCAmelCase__ : Dict = classifier_dropout # additional properties lowerCAmelCase__ : int = max_depth lowerCAmelCase__ : List[Any] = max_xpath_tag_unit_embeddings lowerCAmelCase__ : Optional[Any] = max_xpath_subs_unit_embeddings lowerCAmelCase__ : Tuple = tag_pad_id lowerCAmelCase__ : Any = subs_pad_id lowerCAmelCase__ : Dict = xpath_unit_hidden_size
94
import flax.linen as nn import jax import jax.numpy as jnp class lowerCamelCase__ ( nn.Module): '''simple docstring''' snake_case_ =42 snake_case_ =jnp.floataa def lowerCAmelCase__ (self ) -> List[str]: """simple docstring""" lowerCAmelCase__ : Dict = nn.Conv( self.out_channels ,kernel_size=(3, 3) ,strides=(1, 1) ,padding=((1, 1), (1, 1)) ,dtype=self.dtype ,) def __call__(self ,__lowerCamelCase ) -> Union[str, Any]: """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : Dict = hidden_states.shape lowerCAmelCase__ : Dict = jax.image.resize( __lowerCamelCase ,shape=(batch, height * 2, width * 2, channels) ,method='''nearest''' ,) lowerCAmelCase__ : Dict = self.conv(__lowerCamelCase ) return hidden_states class lowerCamelCase__ ( nn.Module): '''simple docstring''' snake_case_ =42 snake_case_ =jnp.floataa def lowerCAmelCase__ (self ) -> int: """simple docstring""" lowerCAmelCase__ : List[Any] = nn.Conv( self.out_channels ,kernel_size=(3, 3) ,strides=(2, 2) ,padding=((1, 1), (1, 1)) ,dtype=self.dtype ,) def __call__(self ,__lowerCamelCase ) -> Any: """simple docstring""" lowerCAmelCase__ : Optional[int] = self.conv(__lowerCamelCase ) return hidden_states class lowerCamelCase__ ( nn.Module): '''simple docstring''' snake_case_ =42 snake_case_ =None snake_case_ =0.0 snake_case_ =None snake_case_ =jnp.floataa def lowerCAmelCase__ (self ) -> int: """simple docstring""" lowerCAmelCase__ : List[str] = self.in_channels if self.out_channels is None else self.out_channels lowerCAmelCase__ : Union[str, Any] = nn.GroupNorm(num_groups=32 ,epsilon=1e-5 ) lowerCAmelCase__ : Union[str, Any] = nn.Conv( __lowerCamelCase ,kernel_size=(3, 3) ,strides=(1, 1) ,padding=((1, 1), (1, 1)) ,dtype=self.dtype ,) lowerCAmelCase__ : int = nn.Dense(__lowerCamelCase ,dtype=self.dtype ) lowerCAmelCase__ : List[Any] = nn.GroupNorm(num_groups=32 ,epsilon=1e-5 ) lowerCAmelCase__ : Union[str, Any] = nn.Dropout(self.dropout_prob ) lowerCAmelCase__ : Optional[int] = nn.Conv( __lowerCamelCase ,kernel_size=(3, 3) ,strides=(1, 1) ,padding=((1, 1), (1, 1)) ,dtype=self.dtype ,) lowerCAmelCase__ : str = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut lowerCAmelCase__ : Union[str, Any] = None if use_nin_shortcut: lowerCAmelCase__ : Optional[Any] = nn.Conv( __lowerCamelCase ,kernel_size=(1, 1) ,strides=(1, 1) ,padding='''VALID''' ,dtype=self.dtype ,) def __call__(self ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase=True ) -> Union[str, Any]: """simple docstring""" lowerCAmelCase__ : Tuple = hidden_states lowerCAmelCase__ : Dict = self.norma(__lowerCamelCase ) lowerCAmelCase__ : Union[str, Any] = nn.swish(__lowerCamelCase ) lowerCAmelCase__ : List[Any] = self.conva(__lowerCamelCase ) lowerCAmelCase__ : List[Any] = self.time_emb_proj(nn.swish(__lowerCamelCase ) ) lowerCAmelCase__ : Dict = jnp.expand_dims(jnp.expand_dims(__lowerCamelCase ,1 ) ,1 ) lowerCAmelCase__ : Optional[int] = hidden_states + temb lowerCAmelCase__ : Optional[int] = self.norma(__lowerCamelCase ) lowerCAmelCase__ : Optional[int] = nn.swish(__lowerCamelCase ) lowerCAmelCase__ : int = self.dropout(__lowerCamelCase ,__lowerCamelCase ) lowerCAmelCase__ : List[Any] = self.conva(__lowerCamelCase ) if self.conv_shortcut is not None: lowerCAmelCase__ : Optional[int] = self.conv_shortcut(__lowerCamelCase ) return hidden_states + residual
94
1
import unittest import torch from diffusers import VQModel from diffusers.utils import floats_tensor, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class _lowercase ( lowercase__ , lowercase__ , unittest.TestCase): """simple docstring""" A__ = VQModel A__ = "sample" @property def lowerCAmelCase ( self : List[Any] , __lowerCamelCase : Tuple=(32, 32) ): '''simple docstring''' lowerCamelCase__ : Any = 4 lowerCamelCase__ : Tuple = 3 lowerCamelCase__ : Any = floats_tensor((batch_size, num_channels) + sizes ).to(__lowerCamelCase ) return {"sample": image} @property def lowerCAmelCase ( self : Tuple ): '''simple docstring''' return (3, 32, 32) @property def lowerCAmelCase ( self : str ): '''simple docstring''' return (3, 32, 32) def lowerCAmelCase ( self : str ): '''simple docstring''' lowerCamelCase__ : Union[str, Any] = { "block_out_channels": [32, 64], "in_channels": 3, "out_channels": 3, "down_block_types": ["DownEncoderBlock2D", "DownEncoderBlock2D"], "up_block_types": ["UpDecoderBlock2D", "UpDecoderBlock2D"], "latent_channels": 3, } lowerCamelCase__ : Optional[Any] = self.dummy_input return init_dict, inputs_dict def lowerCAmelCase ( self : int ): '''simple docstring''' pass def lowerCAmelCase ( self : Optional[Any] ): '''simple docstring''' pass def lowerCAmelCase ( self : List[Any] ): '''simple docstring''' lowerCamelCase__ , lowerCamelCase__ : str = VQModel.from_pretrained("fusing/vqgan-dummy" , output_loading_info=__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(__lowerCamelCase ) lowerCamelCase__ : Optional[int] = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def lowerCAmelCase ( self : Optional[int] ): '''simple docstring''' lowerCamelCase__ : List[Any] = VQModel.from_pretrained("fusing/vqgan-dummy" ) model.to(__lowerCamelCase ).eval() torch.manual_seed(0 ) if torch.cuda.is_available(): torch.cuda.manual_seed_all(0 ) lowerCamelCase__ : Union[str, Any] = torch.randn(1 , model.config.in_channels , model.config.sample_size , model.config.sample_size ) lowerCamelCase__ : int = image.to(__lowerCamelCase ) with torch.no_grad(): lowerCamelCase__ : Any = model(__lowerCamelCase ).sample lowerCamelCase__ : Dict = output[0, -1, -3:, -3:].flatten().cpu() # fmt: off lowerCamelCase__ : int = torch.tensor([-0.0_1_5_3, -0.4_0_4_4, -0.1_8_8_0, -0.5_1_6_1, -0.2_4_1_8, -0.4_0_7_2, -0.1_6_1_2, -0.0_6_3_3, -0.0_1_4_3] ) # fmt: on self.assertTrue(torch.allclose(__lowerCamelCase , __lowerCamelCase , atol=1E-3 ) )
184
from PIL import Image def lowercase_ ( _A : Image ): """simple docstring""" lowerCamelCase__ , lowerCamelCase__ : List[str] = image.size lowerCamelCase__ : Dict = 0 lowerCamelCase__ : Dict = image.load() for i in range(_A ): for j in range(_A ): lowerCamelCase__ : List[str] = pixels[j, i] mean += pixel mean //= width * height for j in range(_A ): for i in range(_A ): lowerCamelCase__ : Union[str, Any] = 255 if pixels[i, j] > mean else 0 return image if __name__ == "__main__": A : int = mean_threshold(Image.open("path_to_image").convert("L")) image.save("output_image_path")
184
1
"""simple docstring""" import unittest import torch from diffusers import VQModel from diffusers.utils import floats_tensor, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class lowerCamelCase_ ( _a , _a , unittest.TestCase ): SCREAMING_SNAKE_CASE_ = VQModel SCREAMING_SNAKE_CASE_ = """sample""" @property def SCREAMING_SNAKE_CASE_ ( self : Any ,__lowerCamelCase : int=(32, 32) ): '''simple docstring''' a = 4 a = 3 a = floats_tensor((batch_size, num_channels) + sizes ).to(__lowerCamelCase ) return {"sample": image} @property def SCREAMING_SNAKE_CASE_ ( self : Dict ): '''simple docstring''' return (3, 32, 32) @property def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] ): '''simple docstring''' return (3, 32, 32) def SCREAMING_SNAKE_CASE_ ( self : Dict ): '''simple docstring''' a = { """block_out_channels""": [32, 64], """in_channels""": 3, """out_channels""": 3, """down_block_types""": ["""DownEncoderBlock2D""", """DownEncoderBlock2D"""], """up_block_types""": ["""UpDecoderBlock2D""", """UpDecoderBlock2D"""], """latent_channels""": 3, } a = self.dummy_input return init_dict, inputs_dict def SCREAMING_SNAKE_CASE_ ( self : Any ): '''simple docstring''' pass def SCREAMING_SNAKE_CASE_ ( self : str ): '''simple docstring''' pass def SCREAMING_SNAKE_CASE_ ( self : Tuple ): '''simple docstring''' a = VQModel.from_pretrained('''fusing/vqgan-dummy''' ,output_loading_info=__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) self.assertEqual(len(loading_info['''missing_keys'''] ) ,0 ) model.to(__lowerCamelCase ) a = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def SCREAMING_SNAKE_CASE_ ( self : Dict ): '''simple docstring''' a = VQModel.from_pretrained('''fusing/vqgan-dummy''' ) model.to(__lowerCamelCase ).eval() torch.manual_seed(0 ) if torch.cuda.is_available(): torch.cuda.manual_seed_all(0 ) a = torch.randn(1 ,model.config.in_channels ,model.config.sample_size ,model.config.sample_size ) a = image.to(__lowerCamelCase ) with torch.no_grad(): a = model(__lowerCamelCase ).sample a = output[0, -1, -3:, -3:].flatten().cpu() # fmt: off a = torch.tensor([-0.0_153, -0.4_044, -0.1_880, -0.5_161, -0.2_418, -0.4_072, -0.1_612, -0.0_633, -0.0_143] ) # fmt: on self.assertTrue(torch.allclose(__lowerCamelCase ,__lowerCamelCase ,atol=1e-3 ) )
368
from __future__ import annotations from collections.abc import Sequence from typing import Literal def SCREAMING_SNAKE_CASE__ ( snake_case_, snake_case_ ) -> str | Literal[False]: """simple docstring""" a = list(snake_case_ ) a = list(snake_case_ ) a = 0 for i in range(len(snake_case_ ) ): if lista[i] != lista[i]: count += 1 a = '''_''' if count > 1: return False else: return "".join(snake_case_ ) def SCREAMING_SNAKE_CASE__ ( snake_case_ ) -> list[str]: """simple docstring""" a = [] while True: a = ['''$'''] * len(snake_case_ ) a = [] for i in range(len(snake_case_ ) ): for j in range(i + 1, len(snake_case_ ) ): a = compare_string(binary[i], binary[j] ) if k is False: a = '''*''' a = '''*''' temp.append('''X''' ) for i in range(len(snake_case_ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(snake_case_ ) == 0: return pi a = list(set(snake_case_ ) ) def SCREAMING_SNAKE_CASE__ ( snake_case_, snake_case_ ) -> list[str]: """simple docstring""" a = [] for minterm in minterms: a = '''''' for _ in range(snake_case_ ): a = str(minterm % 2 ) + string minterm //= 2 temp.append(snake_case_ ) return temp def SCREAMING_SNAKE_CASE__ ( snake_case_, snake_case_, snake_case_ ) -> bool: """simple docstring""" a = list(snake_case_ ) a = list(snake_case_ ) a = 0 for i in range(len(snake_case_ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def SCREAMING_SNAKE_CASE__ ( snake_case_, snake_case_ ) -> list[str]: """simple docstring""" a = [] a = [0] * len(snake_case_ ) for i in range(len(chart[0] ) ): a = 0 a = -1 for j in range(len(snake_case_ ) ): if chart[j][i] == 1: count += 1 a = j if count == 1: a = 1 for i in range(len(snake_case_ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(snake_case_ ) ): a = 0 temp.append(prime_implicants[i] ) while True: a = 0 a = -1 a = 0 for i in range(len(snake_case_ ) ): a = chart[i].count(1 ) if count_n > max_n: a = count_n a = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(snake_case_ ) ): a = 0 def SCREAMING_SNAKE_CASE__ ( snake_case_, snake_case_ ) -> list[list[int]]: """simple docstring""" a = [[0 for x in range(len(snake_case_ ) )] for x in range(len(snake_case_ ) )] for i in range(len(snake_case_ ) ): a = prime_implicants[i].count('''_''' ) for j in range(len(snake_case_ ) ): if is_for_table(prime_implicants[i], binary[j], snake_case_ ): a = 1 return chart def SCREAMING_SNAKE_CASE__ ( ) -> None: """simple docstring""" a = int(input('''Enter the no. of variables\n''' ) ) a = [ float(snake_case_ ) for x in input( '''Enter the decimal representation of Minterms \'Spaces Separated\'\n''' ).split() ] a = decimal_to_binary(snake_case_, snake_case_ ) a = check(snake_case_ ) print('''Prime Implicants are:''' ) print(snake_case_ ) a = prime_implicant_chart(snake_case_, snake_case_ ) a = selection(snake_case_, snake_case_ ) print('''Essential Prime Implicants are:''' ) print(snake_case_ ) if __name__ == "__main__": import doctest doctest.testmod() main()
330
0
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.speechta import SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaProcessor from ..utils import is_datasets_available from .base import PipelineTool if is_datasets_available(): from datasets import load_dataset class SCREAMING_SNAKE_CASE__ ( lowercase__ ): snake_case__ : int = '''microsoft/speecht5_tts''' snake_case__ : List[Any] = ( '''This is a tool that reads an English text out loud. It takes an input named `text` which should contain the ''' '''text to read (in English) and returns a waveform object containing the sound.''' ) snake_case__ : Optional[Any] = '''text_reader''' snake_case__ : int = SpeechTaProcessor snake_case__ : Any = SpeechTaForTextToSpeech snake_case__ : List[str] = SpeechTaHifiGan snake_case__ : int = ['''text'''] snake_case__ : Dict = ['''audio'''] def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Tuple: if self.post_processor is None: a_ : List[Any] = 'microsoft/speecht5_hifigan' super().setup() def SCREAMING_SNAKE_CASE ( self : Optional[Any] , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : Optional[int]=None ) -> str: a_ : Union[str, Any] = self.pre_processor(text=SCREAMING_SNAKE_CASE__ , return_tensors='pt' , truncation=SCREAMING_SNAKE_CASE__ ) if speaker_embeddings is None: if not is_datasets_available(): raise ImportError('Datasets needs to be installed if not passing speaker embeddings.' ) a_ : str = load_dataset('Matthijs/cmu-arctic-xvectors' , split='validation' ) a_ : List[str] = torch.tensor(embeddings_dataset[7_3_0_5]['xvector'] ).unsqueeze(0 ) return {"input_ids": inputs["input_ids"], "speaker_embeddings": speaker_embeddings} def SCREAMING_SNAKE_CASE ( self : List[Any] , SCREAMING_SNAKE_CASE__ : int ) -> Any: with torch.no_grad(): return self.model.generate_speech(**SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE ( self : Tuple , SCREAMING_SNAKE_CASE__ : List[str] ) -> List[str]: with torch.no_grad(): return self.post_processor(SCREAMING_SNAKE_CASE__ ).cpu().detach()
32
from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def SCREAMING_SNAKE_CASE_ ( ) -> Any: """simple docstring""" a_ : Optional[Any] = HfArgumentParser(__A ) a_ : Optional[int] = parser.parse_args_into_dataclasses()[0] a_ : List[Any] = TensorFlowBenchmark(args=__A ) try: a_ : List[str] = parser.parse_args_into_dataclasses()[0] except ValueError as e: a_ : Dict = 'Arg --no_{0} is no longer used, please use --no-{0} instead.' a_ : Dict = ' '.join(str(__A ).split(' ' )[:-1] ) a_ : int = '' a_ : int = eval(str(__A ).split(' ' )[-1] ) a_ : Any = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:] ) else: wrong_args.append(__A ) if len(__A ) > 0: a_ : str = full_error_msg + begin_error_msg + str(__A ) raise ValueError(__A ) benchmark.run() if __name__ == "__main__": main()
32
1
def UpperCamelCase_( _snake_case : int ): """simple docstring""" __a =abs(_snake_case ) __a =0 while n > 0: res += n % 10 n //= 10 return res def UpperCamelCase_( _snake_case : int ): """simple docstring""" __a =abs(_snake_case ) return n if n < 10 else n % 10 + sum_of_digits(n // 10 ) def UpperCamelCase_( _snake_case : int ): """simple docstring""" return sum(int(_snake_case ) for c in str(abs(_snake_case ) ) ) def UpperCamelCase_( ): """simple docstring""" from collections.abc import Callable from timeit import timeit def benchmark_a_function(_snake_case : Callable , _snake_case : int ) -> None: __a =F'{func.__name__}({value})' __a =timeit(F'__main__.{call}' , setup='import __main__' ) print(F'{call:56} = {func(_snake_case )} -- {timing:.4f} seconds' ) for value in (262144, 1125899906842624, 1267650600228229401496703205376): for func in (sum_of_digits, sum_of_digits_recursion, sum_of_digits_compact): benchmark_a_function(_snake_case , _snake_case ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
308
import string # frequency taken from https://en.wikipedia.org/wiki/Letter_frequency _lowerCAmelCase : Tuple = { "E": 12.70, "T": 9.06, "A": 8.17, "O": 7.51, "I": 6.97, "N": 6.75, "S": 6.33, "H": 6.09, "R": 5.99, "D": 4.25, "L": 4.03, "C": 2.78, "U": 2.76, "M": 2.41, "W": 2.36, "F": 2.23, "G": 2.02, "Y": 1.97, "P": 1.93, "B": 1.29, "V": 0.98, "K": 0.77, "J": 0.15, "X": 0.15, "Q": 0.10, "Z": 0.07, } _lowerCAmelCase : Optional[int] = "ETAOINSHRDLCUMWFGYPBVKJXQZ" _lowerCAmelCase : Any = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" def UpperCamelCase_( _snake_case : str ): """simple docstring""" __a ={letter: 0 for letter in string.ascii_uppercase} for letter in message.upper(): if letter in LETTERS: letter_count[letter] += 1 return letter_count def UpperCamelCase_( _snake_case : tuple ): """simple docstring""" return x[0] def UpperCamelCase_( _snake_case : str ): """simple docstring""" __a =get_letter_count(_snake_case ) __a ={ freq: [] for letter, freq in letter_to_freq.items() } for letter in LETTERS: freq_to_letter[letter_to_freq[letter]].append(_snake_case ) __a ={} for freq in freq_to_letter: freq_to_letter[freq].sort(key=ETAOIN.find , reverse=_snake_case ) __a =''.join(freq_to_letter[freq] ) __a =list(freq_to_letter_str.items() ) freq_pairs.sort(key=_snake_case , reverse=_snake_case ) __a =[freq_pair[1] for freq_pair in freq_pairs] return "".join(_snake_case ) def UpperCamelCase_( _snake_case : str ): """simple docstring""" __a =get_frequency_order(_snake_case ) __a =0 for common_letter in ETAOIN[:6]: if common_letter in freq_order[:6]: match_score += 1 for uncommon_letter in ETAOIN[-6:]: if uncommon_letter in freq_order[-6:]: match_score += 1 return match_score if __name__ == "__main__": import doctest doctest.testmod()
308
1
'''simple docstring''' import warnings from diffusers import StableDiffusionImgaImgPipeline # noqa F401 warnings.warn( 'The `image_to_image.py` script is outdated. Please use directly `from diffusers import' ' StableDiffusionImg2ImgPipeline` instead.' )
234
'''simple docstring''' import random def lowercase__ ( __lowercase : list , __lowercase : Optional[Any] ) -> tuple: """simple docstring""" __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = [], [], [] for element in data: if element < pivot: less.append(__lowercase ) elif element > pivot: greater.append(__lowercase ) else: equal.append(__lowercase ) return less, equal, greater def lowercase__ ( __lowercase : list , __lowercase : int ) -> Dict: """simple docstring""" if index >= len(__lowercase ) or index < 0: return None __UpperCamelCase = items[random.randint(0 , len(__lowercase ) - 1 )] __UpperCamelCase = 0 __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = _partition(__lowercase , __lowercase ) __UpperCamelCase = len(__lowercase ) __UpperCamelCase = len(__lowercase ) # index is the pivot if m <= index < m + count: return pivot # must be in smaller elif m > index: return quick_select(__lowercase , __lowercase ) # must be in larger else: return quick_select(__lowercase , index - (m + count) )
53
0
import timeit import numpy as np import datasets from datasets.arrow_writer import ArrowWriter from datasets.features.features import _ArrayXD def __UpperCamelCase (_SCREAMING_SNAKE_CASE ) -> Any: def wrapper(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): lowercase__ = timeit.default_timer() lowercase__ = func(*_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) lowercase__ = timeit.default_timer() - starttime return delta lowercase__ = func.__name__ return wrapper def __UpperCamelCase (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=100 , _SCREAMING_SNAKE_CASE=None ) -> int: lowercase__ = [] lowercase__ = seq_shapes or {} for i in range(_SCREAMING_SNAKE_CASE ): lowercase__ = {} for col_id, (k, v) in enumerate(features.items() ): if isinstance(_SCREAMING_SNAKE_CASE , _ArrayXD ): lowercase__ = np.random.rand(*v.shape ).astype(v.dtype ) elif isinstance(_SCREAMING_SNAKE_CASE , datasets.Value ): if v.dtype == "string": lowercase__ = 'The small grey turtle was surprisingly fast when challenged.' else: lowercase__ = np.random.randint(10 , size=1 ).astype(v.dtype ).item() elif isinstance(_SCREAMING_SNAKE_CASE , datasets.Sequence ): while isinstance(_SCREAMING_SNAKE_CASE , datasets.Sequence ): lowercase__ = v.feature lowercase__ = seq_shapes[k] lowercase__ = np.random.rand(*_SCREAMING_SNAKE_CASE ).astype(v.dtype ) lowercase__ = data dummy_data.append((i, example) ) return dummy_data def __UpperCamelCase (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=100 , _SCREAMING_SNAKE_CASE=None ) -> Any: lowercase__ = generate_examples(_SCREAMING_SNAKE_CASE , num_examples=_SCREAMING_SNAKE_CASE , seq_shapes=_SCREAMING_SNAKE_CASE ) with ArrowWriter(features=_SCREAMING_SNAKE_CASE , path=_SCREAMING_SNAKE_CASE ) as writer: for key, record in dummy_data: lowercase__ = features.encode_example(_SCREAMING_SNAKE_CASE ) writer.write(_SCREAMING_SNAKE_CASE ) lowercase__ , lowercase__ = writer.finalize() if not num_final_examples == num_examples: raise ValueError( F"""Error writing the dataset, wrote {num_final_examples} examples but should have written {num_examples}.""" ) lowercase__ = datasets.Dataset.from_file(filename=_SCREAMING_SNAKE_CASE , info=datasets.DatasetInfo(features=_SCREAMING_SNAKE_CASE ) ) return dataset
269
from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class SCREAMING_SNAKE_CASE (UpperCAmelCase ): def __init__( self : List[str] , a : Callable , a : Optional[Features] = None , a : str = None , a : bool = False , a : bool = False , a : Optional[dict] = None , a : Optional[int] = None , **a : str , )-> Tuple: """simple docstring""" super().__init__( features=a , cache_dir=a , keep_in_memory=a , streaming=a , num_proc=a , **a , ) lowercase__ = Generator( cache_dir=a , features=a , generator=a , gen_kwargs=a , **a , ) def SCREAMING_SNAKE_CASE_ ( self : Any )-> Dict: """simple docstring""" if self.streaming: lowercase__ = self.builder.as_streaming_dataset(split='train' ) # Build regular (map-style) dataset else: lowercase__ = None lowercase__ = None lowercase__ = None lowercase__ = None self.builder.download_and_prepare( download_config=a , download_mode=a , verification_mode=a , base_path=a , num_proc=self.num_proc , ) lowercase__ = self.builder.as_dataset( split='train' , verification_mode=a , in_memory=self.keep_in_memory ) return dataset
269
1
__A : int = '2.13.1' import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse('3.7'): raise ImportWarning( 'To use `datasets`, Python>=3.7 is required, and the current version of Python doesn\'t match this condition.' ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( 'To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn\'t match this condition.\n' 'If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.' ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip __A : Dict = concatenate_datasets __A : Any = DownloadConfig __A : Dict = DownloadManager __A : Dict = DownloadMode __A : Dict = DownloadConfig __A : Optional[int] = DownloadMode __A : List[Any] = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
154
"""simple docstring""" from decimal import Decimal, getcontext from math import ceil, factorial def __lowerCamelCase ( __UpperCamelCase ) -> str: """simple docstring""" if not isinstance(__UpperCamelCase , __UpperCamelCase ): raise TypeError("Undefined for non-integers" ) elif precision < 1: raise ValueError("Undefined for non-natural numbers" ) lowerCAmelCase_ : str = precision lowerCAmelCase_ : Tuple = ceil(precision / 14 ) lowerCAmelCase_ : Optional[int] = 426880 * Decimal(10005 ).sqrt() lowerCAmelCase_ : Any = 1 lowerCAmelCase_ : Any = 13591409 lowerCAmelCase_ : List[str] = Decimal(__UpperCamelCase ) for k in range(1 , __UpperCamelCase ): lowerCAmelCase_ : List[Any] = factorial(6 * k ) // (factorial(3 * k ) * factorial(__UpperCamelCase ) ** 3) linear_term += 545140134 exponential_term *= -262537412640768000 partial_sum += Decimal(multinomial_term * linear_term ) / exponential_term return str(constant_term / partial_sum )[:-1] if __name__ == "__main__": lowercase__ = 50 print(F"""The first {n} digits of pi is: {pi(n)}""")
241
0
"""simple docstring""" from manim import * class __magic_name__ ( __UpperCAmelCase ): def __snake_case ( self : Optional[int] ): '''simple docstring''' lowercase :Optional[int] = Rectangle(height=0.5 , width=0.5 ) lowercase :int = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) lowercase :Optional[int] = [mem.copy() for i in range(6 )] lowercase :List[str] = [mem.copy() for i in range(6 )] lowercase :List[Any] = VGroup(*snake_case__ ).arrange(snake_case__ , buff=0 ) lowercase :List[Any] = VGroup(*snake_case__ ).arrange(snake_case__ , buff=0 ) lowercase :Union[str, Any] = VGroup(snake_case__ , snake_case__ ).arrange(snake_case__ , buff=0 ) lowercase :Dict = Text('''CPU''' , font_size=2_4 ) lowercase :int = Group(snake_case__ , snake_case__ ).arrange(snake_case__ , buff=0.5 , aligned_edge=snake_case__ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(snake_case__ ) lowercase :Optional[Any] = [mem.copy() for i in range(4 )] lowercase :int = VGroup(*snake_case__ ).arrange(snake_case__ , buff=0 ) lowercase :Tuple = Text('''GPU''' , font_size=2_4 ) lowercase :Any = Group(snake_case__ , snake_case__ ).arrange(snake_case__ , buff=0.5 , aligned_edge=snake_case__ ) gpu.move_to([-1, -1, 0] ) self.add(snake_case__ ) lowercase :List[Any] = [mem.copy() for i in range(6 )] lowercase :Optional[int] = VGroup(*snake_case__ ).arrange(snake_case__ , buff=0 ) lowercase :int = Text('''Model''' , font_size=2_4 ) lowercase :Any = Group(snake_case__ , snake_case__ ).arrange(snake_case__ , buff=0.5 , aligned_edge=snake_case__ ) model.move_to([3, -1.0, 0] ) self.add(snake_case__ ) lowercase :Tuple = [] for i, rect in enumerate(snake_case__ ): rect.set_stroke(snake_case__ ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) lowercase :Optional[int] = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(snake_case__ , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=snake_case__ ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] , direction=snake_case__ , buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] , direction=snake_case__ , buff=0.0 ) self.add(snake_case__ ) cpu_targs.append(snake_case__ ) lowercase :List[str] = [mem.copy() for i in range(6 )] lowercase :List[Any] = VGroup(*snake_case__ ).arrange(snake_case__ , buff=0 ) lowercase :Optional[int] = Text('''Loaded Checkpoint''' , font_size=2_4 ) lowercase :List[Any] = Group(snake_case__ , snake_case__ ).arrange(snake_case__ , aligned_edge=snake_case__ , buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) lowercase :List[Any] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) lowercase :int = MarkupText( f"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=1_8 , ) key_text.move_to([-5, 2.4, 0] ) self.add(snake_case__ , snake_case__ ) lowercase :int = MarkupText( f"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=1_8 , ) blue_text.next_to(snake_case__ , DOWN * 2.4 , aligned_edge=key_text.get_left() ) lowercase :List[str] = MarkupText( f"""Next, a <i><span fgcolor=\"{BLUE}\">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor=\"{BLUE}\">single shard</span>.""" , font_size=2_4 , ) step_a.move_to([2, 2, 0] ) self.play(Write(snake_case__ ) , Write(snake_case__ ) ) self.play(Write(snake_case__ , run_time=1 ) , Create(snake_case__ , run_time=1 ) ) lowercase :str = [] lowercase :List[Any] = [] for i, rect in enumerate(snake_case__ ): lowercase :Tuple = fill.copy().set_fill(snake_case__ , opacity=0.7 ) target.move_to(snake_case__ ) first_animations.append(GrowFromCenter(snake_case__ , run_time=1 ) ) lowercase :Tuple = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(snake_case__ , run_time=1.5 ) ) self.play(*snake_case__ ) self.play(*snake_case__ ) self.wait()
172
"""simple docstring""" def lowerCamelCase (a_ :list , a_ :list , a_ :int , a_ :int , a_ :int) -> int: if index == number_of_items: return 0 lowercase :Optional[int] = 0 lowercase :str = 0 lowercase :List[str] = knapsack(a_ , a_ , a_ , a_ , index + 1) if weights[index] <= max_weight: lowercase :Any = values[index] + knapsack( a_ , a_ , a_ , max_weight - weights[index] , index + 1) return max(a_ , a_) if __name__ == "__main__": import doctest doctest.testmod()
172
1
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase_ : Union[str, Any] = logging.get_logger(__name__) lowerCAmelCase_ : Optional[Any] = { 'roberta-base': 'https://huggingface.co/roberta-base/resolve/main/config.json', 'roberta-large': 'https://huggingface.co/roberta-large/resolve/main/config.json', 'roberta-large-mnli': 'https://huggingface.co/roberta-large-mnli/resolve/main/config.json', 'distilroberta-base': 'https://huggingface.co/distilroberta-base/resolve/main/config.json', 'roberta-base-openai-detector': 'https://huggingface.co/roberta-base-openai-detector/resolve/main/config.json', 'roberta-large-openai-detector': 'https://huggingface.co/roberta-large-openai-detector/resolve/main/config.json', } class __SCREAMING_SNAKE_CASE (lowerCamelCase_ ): """simple docstring""" __a ='roberta' def __init__( self : Any , __a : List[Any]=5_02_65 , __a : Union[str, Any]=7_68 , __a : Optional[Any]=12 , __a : Any=12 , __a : Union[str, Any]=30_72 , __a : List[str]="gelu" , __a : Optional[int]=0.1 , __a : Optional[Any]=0.1 , __a : Optional[int]=5_12 , __a : Any=2 , __a : Dict=0.02 , __a : List[Any]=1e-1_2 , __a : Any=1 , __a : Optional[int]=0 , __a : int=2 , __a : List[str]="absolute" , __a : Dict=True , __a : Optional[int]=None , **__a : List[str] , ): super().__init__(pad_token_id=__a , bos_token_id=__a , eos_token_id=__a , **__a ) _a = vocab_size _a = hidden_size _a = num_hidden_layers _a = num_attention_heads _a = hidden_act _a = intermediate_size _a = hidden_dropout_prob _a = attention_probs_dropout_prob _a = max_position_embeddings _a = type_vocab_size _a = initializer_range _a = layer_norm_eps _a = position_embedding_type _a = use_cache _a = classifier_dropout class __SCREAMING_SNAKE_CASE (lowerCamelCase_ ): """simple docstring""" @property def UpperCamelCase__ ( self : Tuple ): if self.task == "multiple-choice": _a = {0: "batch", 1: "choice", 2: "sequence"} else: _a = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
63
import argparse from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection from diffusers import UnCLIPImageVariationPipeline, UnCLIPPipeline if __name__ == "__main__": UpperCAmelCase : Dict =argparse.ArgumentParser() parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the output model.""") parser.add_argument( """--txt2img_unclip""", default="""kakaobrain/karlo-v1-alpha""", type=str, required=False, help="""The pretrained txt2img unclip.""", ) UpperCAmelCase : Optional[int] =parser.parse_args() UpperCAmelCase : List[Any] =UnCLIPPipeline.from_pretrained(args.txtaimg_unclip) UpperCAmelCase : List[str] =CLIPImageProcessor() UpperCAmelCase : Optional[int] =CLIPVisionModelWithProjection.from_pretrained("""openai/clip-vit-large-patch14""") UpperCAmelCase : Any =UnCLIPImageVariationPipeline( decoder=txtaimg.decoder, text_encoder=txtaimg.text_encoder, tokenizer=txtaimg.tokenizer, text_proj=txtaimg.text_proj, feature_extractor=feature_extractor, image_encoder=image_encoder, super_res_first=txtaimg.super_res_first, super_res_last=txtaimg.super_res_last, decoder_scheduler=txtaimg.decoder_scheduler, super_res_scheduler=txtaimg.super_res_scheduler, ) imgaimg.save_pretrained(args.dump_path)
128
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { 'microsoft/beit-base-patch16-224-pt22k': ( 'https://huggingface.co/microsoft/beit-base-patch16-224-pt22k/resolve/main/config.json' ), # See all BEiT models at https://huggingface.co/models?filter=beit } class __SCREAMING_SNAKE_CASE ( lowerCamelCase ): snake_case_ = """beit""" def __init__( self : Dict , __lowercase : List[str]=81_92 , __lowercase : str=7_68 , __lowercase : Optional[Any]=12 , __lowercase : Dict=12 , __lowercase : List[str]=30_72 , __lowercase : Dict="gelu" , __lowercase : List[Any]=0.0 , __lowercase : Any=0.0 , __lowercase : str=0.02 , __lowercase : Union[str, Any]=1e-12 , __lowercase : Dict=2_24 , __lowercase : List[Any]=16 , __lowercase : Dict=3 , __lowercase : str=False , __lowercase : Optional[int]=False , __lowercase : Union[str, Any]=False , __lowercase : List[str]=False , __lowercase : int=0.1 , __lowercase : Dict=0.1 , __lowercase : List[str]=True , __lowercase : int=[3, 5, 7, 11] , __lowercase : int=[1, 2, 3, 6] , __lowercase : List[str]=True , __lowercase : Optional[Any]=0.4 , __lowercase : List[Any]=2_56 , __lowercase : Dict=1 , __lowercase : Tuple=False , __lowercase : Optional[int]=2_55 , **__lowercase : Optional[int] , ) -> Optional[Any]: super().__init__(**__lowercase ) SCREAMING_SNAKE_CASE__ : str =vocab_size SCREAMING_SNAKE_CASE__ : Any =hidden_size SCREAMING_SNAKE_CASE__ : Any =num_hidden_layers SCREAMING_SNAKE_CASE__ : List[Any] =num_attention_heads SCREAMING_SNAKE_CASE__ : Optional[Any] =intermediate_size SCREAMING_SNAKE_CASE__ : Union[str, Any] =hidden_act SCREAMING_SNAKE_CASE__ : Optional[Any] =hidden_dropout_prob SCREAMING_SNAKE_CASE__ : Optional[Any] =attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ : Tuple =initializer_range SCREAMING_SNAKE_CASE__ : Dict =layer_norm_eps SCREAMING_SNAKE_CASE__ : Dict =image_size SCREAMING_SNAKE_CASE__ : Optional[Any] =patch_size SCREAMING_SNAKE_CASE__ : Optional[int] =num_channels SCREAMING_SNAKE_CASE__ : int =use_mask_token SCREAMING_SNAKE_CASE__ : Optional[Any] =use_absolute_position_embeddings SCREAMING_SNAKE_CASE__ : List[Any] =use_relative_position_bias SCREAMING_SNAKE_CASE__ : List[Any] =use_shared_relative_position_bias SCREAMING_SNAKE_CASE__ : Dict =layer_scale_init_value SCREAMING_SNAKE_CASE__ : List[Any] =drop_path_rate SCREAMING_SNAKE_CASE__ : int =use_mean_pooling # decode head attributes (semantic segmentation) SCREAMING_SNAKE_CASE__ : Optional[Any] =out_indices SCREAMING_SNAKE_CASE__ : List[str] =pool_scales # auxiliary head attributes (semantic segmentation) SCREAMING_SNAKE_CASE__ : List[Any] =use_auxiliary_head SCREAMING_SNAKE_CASE__ : Optional[int] =auxiliary_loss_weight SCREAMING_SNAKE_CASE__ : str =auxiliary_channels SCREAMING_SNAKE_CASE__ : List[Any] =auxiliary_num_convs SCREAMING_SNAKE_CASE__ : Tuple =auxiliary_concat_input SCREAMING_SNAKE_CASE__ : List[str] =semantic_loss_ignore_index class __SCREAMING_SNAKE_CASE ( lowerCamelCase ): snake_case_ = version.parse("""1.11""" ) @property def __magic_name__ ( self : Optional[int] ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def __magic_name__ ( self : Union[str, Any] ) -> float: return 1e-4
222
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class __SCREAMING_SNAKE_CASE ( lowerCamelCase ): snake_case_ = """facebook/bart-large-mnli""" snake_case_ = ( """This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which """ """should be the text to classify, and `labels`, which should be the list of labels to use for classification. """ """It returns the most likely label in the list of provided `labels` for the input text.""" ) snake_case_ = """text_classifier""" snake_case_ = AutoTokenizer snake_case_ = AutoModelForSequenceClassification snake_case_ = ["""text""", ["""text"""]] snake_case_ = ["""text"""] def __magic_name__ ( self : Any ) -> Any: super().setup() SCREAMING_SNAKE_CASE__ : int =self.model.config SCREAMING_SNAKE_CASE__ : int =-1 for idx, label in config.idalabel.items(): if label.lower().startswith('''entail''' ): SCREAMING_SNAKE_CASE__ : Any =int(__lowercase ) if self.entailment_id == -1: raise ValueError('''Could not determine the entailment ID from the model config, please pass it at init.''' ) def __magic_name__ ( self : str , __lowercase : Any , __lowercase : Optional[Any] ) -> Optional[Any]: SCREAMING_SNAKE_CASE__ : Dict =labels return self.pre_processor( [text] * len(__lowercase ) , [F"This example is {label}" for label in labels] , return_tensors='''pt''' , padding='''max_length''' , ) def __magic_name__ ( self : List[Any] , __lowercase : str ) -> List[Any]: SCREAMING_SNAKE_CASE__ : int =outputs.logits SCREAMING_SNAKE_CASE__ : Any =torch.argmax(logits[:, 2] ).item() return self._labels[label_id]
222
1
import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeqaSeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging SCREAMING_SNAKE_CASE :str = logging.get_logger(__name__) def UpperCAmelCase ( a_ , a_ , a_ ) -> None: """simple docstring""" __A = nn.ModuleList([src_layers[i] for i in layers_to_copy] ) assert len(a_ ) == len(a_ ), F'''{len(a_ )} != {len(a_ )}''' dest_layers.load_state_dict(layers_to_copy.state_dict() ) SCREAMING_SNAKE_CASE :Tuple = { # maps num layers in teacher -> num_layers in student -> which teacher layers to copy. # 12: bart, 16: pegasus, 6: marian/Helsinki-NLP 12: { 1: [0], # This says that if the teacher has 12 layers and the student has 1, copy layer 0 of the teacher 2: [0, 6], 3: [0, 6, 11], 4: [0, 4, 8, 11], 6: [0, 2, 4, 7, 9, 11], 9: [0, 1, 2, 4, 5, 7, 9, 10, 11], 12: list(range(12)), }, 16: { # maps num layers in student -> which teacher layers to copy 1: [0], 2: [0, 15], 3: [0, 8, 15], 4: [0, 5, 10, 15], 6: [0, 3, 6, 9, 12, 15], 8: [0, 2, 4, 6, 8, 10, 12, 15], 9: [0, 1, 3, 5, 7, 9, 11, 13, 15], 12: [0, 1, 2, 3, 4, 5, 6, 7, 9, 11, 13, 15], 16: list(range(16)), }, 6: {1: [0], 2: [0, 5], 3: [0, 2, 5], 4: [0, 1, 3, 5], 6: list(range(6))}, } SCREAMING_SNAKE_CASE :List[Any] = { # maps num layers in student -> which teacher layers to copy. 6: {1: [5], 2: [3, 5], 3: [1, 4, 5], 4: [1, 2, 4, 5]}, 12: {1: [11], 2: [5, 11], 3: [3, 7, 11], 6: [1, 3, 5, 8, 10, 11]}, 16: {1: [15], 4: [4, 9, 12, 15], 8: [1, 3, 5, 7, 9, 11, 13, 15]}, } def UpperCAmelCase ( a_ , a_ ) -> str: """simple docstring""" try: __A = LAYERS_TO_COPY[n_teacher][n_student] return val except KeyError: if n_student != n_teacher: warnings.warn( F'''no hardcoded layers to copy for teacher {n_teacher} -> student {n_student}, defaulting to first''' F''' {n_student}''' ) return list(range(a_ ) ) def UpperCAmelCase ( a_ , a_ ) -> List[int]: """simple docstring""" if n_student > n_teacher: raise ValueError(F'''Cannot perform intermediate supervision for student {n_student} > teacher {n_teacher}''' ) elif n_teacher == n_student: return list(range(a_ ) ) elif n_student == 1: return [n_teacher - 1] else: return LAYERS_TO_SUPERVISE[n_teacher][n_student] def UpperCAmelCase ( a_ , a_ = "student" , a_ = None , a_ = None , a_=False , a_=None , a_=None , **a_ , ) -> Tuple[PreTrainedModel, List[int], List[int]]: """simple docstring""" __A = "encoder_layers and decoder_layers cannot be both None-- you would just have an identical teacher." assert (e is not None) or (d is not None), _msg if isinstance(a_ , a_ ): AutoTokenizer.from_pretrained(a_ ).save_pretrained(a_ ) # purely for convenience __A = AutoModelForSeqaSeqLM.from_pretrained(a_ ).eval() else: assert isinstance(a_ , a_ ), F'''teacher must be a model or string got type {type(a_ )}''' __A = teacher.config.to_diff_dict() try: __A , __A = teacher.config.encoder_layers, teacher.config.decoder_layers if e is None: __A = teacher_e if d is None: __A = teacher_d init_kwargs.update({"encoder_layers": e, "decoder_layers": d} ) except AttributeError: # T5 if hasattr(teacher.config , "num_encoder_layers" ): __A , __A = teacher.config.num_encoder_layers, teacher.config.num_decoder_layers else: __A , __A = teacher.config.num_layers, teacher.config.num_decoder_layers if e is None: __A = teacher_e if d is None: __A = teacher_d if hasattr(teacher.config , "num_encoder_layers" ): init_kwargs.update({"num_encoder_layers": e, "num_decoder_layers": d} ) else: init_kwargs.update({"num_layers": e, "num_decoder_layers": d} ) # Kwargs to instantiate student: teacher kwargs with updated layer numbers + **extra_config_kwargs init_kwargs.update(a_ ) # Copy weights __A = teacher.config_class(**a_ ) __A = AutoModelForSeqaSeqLM.from_config(a_ ) # Start by copying the full teacher state dict this will copy the first N teacher layers to the student. __A = student.load_state_dict(teacher.state_dict() , strict=a_ ) assert info.missing_keys == [], info.missing_keys # every student key should have a teacher keys. if copy_first_teacher_layers: # Our copying is done. We just log and save __A , __A = list(range(a_ ) ), list(range(a_ ) ) logger.info( F'''Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to''' F''' {save_path}''' ) student.save_pretrained(a_ ) return student, e_layers_to_copy, d_layers_to_copy # Decide which layers of the teacher to copy. Not exactly alternating -- we try to keep first and last layer. if e_layers_to_copy is None: __A = pick_layers_to_copy(a_ , a_ ) if d_layers_to_copy is None: __A = pick_layers_to_copy(a_ , a_ ) try: if hasattr( a_ , "prophetnet" ): # For ProphetNet, student.model.encoder.layers is called student.prophetnet.encoder.layers copy_layers(teacher.prophetnet.encoder.layers , student.prophetnet.encoder.layers , a_ ) copy_layers(teacher.prophetnet.decoder.layers , student.prophetnet.decoder.layers , a_ ) else: copy_layers(teacher.model.encoder.layers , student.model.encoder.layers , a_ ) copy_layers(teacher.model.decoder.layers , student.model.decoder.layers , a_ ) except AttributeError: # For t5, student.model.encoder.layers is called student.encoder.block copy_layers(teacher.encoder.block , student.encoder.block , a_ ) copy_layers(teacher.decoder.block , student.decoder.block , a_ ) logger.info( F'''Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to {save_path}''' ) __A = { "teacher_type": teacher.config.model_type, "copied_encoder_layers": e_layers_to_copy, "copied_decoder_layers": d_layers_to_copy, } student.save_pretrained(a_ ) # Save information about copying for easier reproducibility return student, e_layers_to_copy, d_layers_to_copy if __name__ == "__main__": fire.Fire(create_student_by_copying_alternating_layers)
15
"""simple docstring""" import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def __magic_name__ ( *lowercase ): if not isinstance(lowercase , lowercase ): SCREAMING_SNAKE_CASE_: Optional[Any] =list(lowercase ) for i in range(len(lowercase ) ): SCREAMING_SNAKE_CASE_: Optional[Any] =None gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() return objects def __magic_name__ ( lowercase ): SCREAMING_SNAKE_CASE_: List[Any] =[ """CUDA out of memory.""", # CUDA OOM """cuDNN error: CUDNN_STATUS_NOT_SUPPORTED.""", # CUDNN SNAFU """DefaultCPUAllocator: can't allocate memory""", # CPU OOM ] if isinstance(lowercase , lowercase ) and len(exception.args ) == 1: return any(err in exception.args[0] for err in _statements ) return False def __magic_name__ ( lowercase = None , lowercase = 128 ): if function is None: return functools.partial(lowercase , starting_batch_size=lowercase ) SCREAMING_SNAKE_CASE_: str =starting_batch_size def decorator(*lowercase , **lowercase ): nonlocal batch_size gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() SCREAMING_SNAKE_CASE_: Optional[int] =list(inspect.signature(lowercase ).parameters.keys() ) # Guard against user error if len(lowercase ) < (len(lowercase ) + 1): SCREAMING_SNAKE_CASE_: List[Any] =""", """.join([f'''{arg}={value}''' for arg, value in zip(params[1:] , args[1:] )] ) raise TypeError( f'''Batch size was passed into `{function.__name__}` as the first argument when called.''' f'''Remove this as the decorator already does so: `{function.__name__}({arg_str})`''' ) while True: if batch_size == 0: raise RuntimeError("""No executable batch size found, reached zero.""" ) try: return function(lowercase , *lowercase , **lowercase ) except Exception as e: if should_reduce_batch_size(lowercase ): gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() batch_size //= 2 else: raise return decorator
173
0
from __future__ import annotations def __lowerCamelCase ( __a :list[list[int]] ): """simple docstring""" for i in range(1 , len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1 , len(__SCREAMING_SNAKE_CASE ) ): matrix[i][0] += matrix[i - 1][0] # updating the path cost for current position for i in range(1 , len(__SCREAMING_SNAKE_CASE ) ): for j in range(1 , len(matrix[0] ) ): matrix[i][j] += min(matrix[i - 1][j] , matrix[i][j - 1] ) return matrix[-1][-1] if __name__ == "__main__": import doctest doctest.testmod()
370
from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging A : Optional[Any] = logging.get_logger(__name__) A : List[str] = { '''huggingface/time-series-transformer-tourism-monthly''': ( '''https://huggingface.co/huggingface/time-series-transformer-tourism-monthly/resolve/main/config.json''' ), # See all TimeSeriesTransformer models at https://huggingface.co/models?filter=time_series_transformer } class A (SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCamelCase : Any = '''time_series_transformer''' __lowerCamelCase : Dict = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', '''num_hidden_layers''': '''encoder_layers''', } def __init__( self : Union[str, Any] , __lowerCAmelCase : Optional[int] = None , __lowerCAmelCase : Optional[int] = None , __lowerCAmelCase : str = "student_t" , __lowerCAmelCase : str = "nll" , __lowerCAmelCase : int = 1 , __lowerCAmelCase : List[int] = [1, 2, 3, 4, 5, 6, 7] , __lowerCAmelCase : Optional[Union[str, bool]] = "mean" , __lowerCAmelCase : int = 0 , __lowerCAmelCase : int = 0 , __lowerCAmelCase : int = 0 , __lowerCAmelCase : int = 0 , __lowerCAmelCase : Optional[List[int]] = None , __lowerCAmelCase : Optional[List[int]] = None , __lowerCAmelCase : int = 32 , __lowerCAmelCase : int = 32 , __lowerCAmelCase : int = 2 , __lowerCAmelCase : int = 2 , __lowerCAmelCase : int = 2 , __lowerCAmelCase : int = 2 , __lowerCAmelCase : bool = True , __lowerCAmelCase : str = "gelu" , __lowerCAmelCase : int = 64 , __lowerCAmelCase : float = 0.1 , __lowerCAmelCase : float = 0.1 , __lowerCAmelCase : float = 0.1 , __lowerCAmelCase : float = 0.1 , __lowerCAmelCase : float = 0.1 , __lowerCAmelCase : int = 1_00 , __lowerCAmelCase : float = 0.0_2 , __lowerCAmelCase : Optional[Any]=True , **__lowerCAmelCase : Union[str, Any] , ) -> Optional[int]: """simple docstring""" A__ = prediction_length A__ = context_length or prediction_length A__ = distribution_output A__ = loss A__ = input_size A__ = num_time_features A__ = lags_sequence A__ = scaling A__ = num_dynamic_real_features A__ = num_static_real_features A__ = num_static_categorical_features if cardinality and num_static_categorical_features > 0: if len(__lowerCAmelCase ) != num_static_categorical_features: raise ValueError( """The cardinality should be a list of the same length as `num_static_categorical_features`""" ) A__ = cardinality else: A__ = [0] if embedding_dimension and num_static_categorical_features > 0: if len(__lowerCAmelCase ) != num_static_categorical_features: raise ValueError( """The embedding dimension should be a list of the same length as `num_static_categorical_features`""" ) A__ = embedding_dimension else: A__ = [min(50 , (cat + 1) // 2 ) for cat in self.cardinality] A__ = num_parallel_samples # Transformer architecture configuration A__ = input_size * len(__lowerCAmelCase ) + self._number_of_features A__ = d_model A__ = encoder_attention_heads A__ = decoder_attention_heads A__ = encoder_ffn_dim A__ = decoder_ffn_dim A__ = encoder_layers A__ = decoder_layers A__ = dropout A__ = attention_dropout A__ = activation_dropout A__ = encoder_layerdrop A__ = decoder_layerdrop A__ = activation_function A__ = init_std A__ = use_cache super().__init__(is_encoder_decoder=__lowerCAmelCase , **__lowerCAmelCase ) @property def a_ ( self : List[Any] ) -> int: """simple docstring""" return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
276
0
from __future__ import annotations import unittest from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotForConditionalGeneration, TFBlenderbotModel @require_tf class _lowercase : """simple docstring""" __A = BlenderbotConfig __A = {} __A = "gelu" def __init__(self , lowerCamelCase_ , lowerCamelCase_=13 , lowerCamelCase_=7 , lowerCamelCase_=True , lowerCamelCase_=False , lowerCamelCase_=99 , lowerCamelCase_=32 , lowerCamelCase_=2 , lowerCamelCase_=4 , lowerCamelCase_=37 , lowerCamelCase_=0.1 , lowerCamelCase_=0.1 , lowerCamelCase_=20 , lowerCamelCase_=2 , lowerCamelCase_=1 , lowerCamelCase_=0 , ): """simple docstring""" a = parent a = batch_size a = seq_length a = is_training a = use_labels a = vocab_size a = hidden_size a = num_hidden_layers a = num_attention_heads a = intermediate_size a = hidden_dropout_prob a = attention_probs_dropout_prob a = max_position_embeddings a = eos_token_id a = pad_token_id a = bos_token_id def UpperCamelCase_ (self ): """simple docstring""" a = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) a = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) a = tf.concat([input_ids, eos_tensor] , axis=1 ) a = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) a = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) a = prepare_blenderbot_inputs_dict(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) return config, inputs_dict def UpperCamelCase_ (self , lowerCamelCase_ , lowerCamelCase_ ): """simple docstring""" a = TFBlenderbotModel(config=lowerCamelCase_ ).get_decoder() a = inputs_dict["input_ids"] a = input_ids[:1, :] a = inputs_dict["attention_mask"][:1, :] a = inputs_dict["head_mask"] a = 1 # first forward pass a = model(lowerCamelCase_ , attention_mask=lowerCamelCase_ , head_mask=lowerCamelCase_ , use_cache=lowerCamelCase_ ) a , a = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids a = ids_tensor((self.batch_size, 3) , config.vocab_size ) a = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and a = tf.concat([input_ids, next_tokens] , axis=-1 ) a = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) a = model(lowerCamelCase_ , attention_mask=lowerCamelCase_ )[0] a = model(lowerCamelCase_ , attention_mask=lowerCamelCase_ , past_key_values=lowerCamelCase_ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice a = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) a = output_from_no_past[:, -3:, random_slice_idx] a = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(lowerCamelCase_ , lowerCamelCase_ , rtol=1E-3 ) def a( A : Optional[Any] , A : int , A : Optional[Any] , A : Tuple=None , A : str=None , A : int=None , A : Dict=None , A : Union[str, Any]=None , ) -> Optional[int]: """simple docstring""" if attention_mask is None: a = tf.cast(tf.math.not_equal(A , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: a = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: a = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: a = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: a = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class _lowercase ( lowerCAmelCase, lowerCAmelCase, unittest.TestCase ): """simple docstring""" __A = (TFBlenderbotForConditionalGeneration, TFBlenderbotModel) if is_tf_available() else () __A = (TFBlenderbotForConditionalGeneration,) if is_tf_available() else () __A = ( { "conversational": TFBlenderbotForConditionalGeneration, "feature-extraction": TFBlenderbotModel, "summarization": TFBlenderbotForConditionalGeneration, "text2text-generation": TFBlenderbotForConditionalGeneration, "translation": TFBlenderbotForConditionalGeneration, } if is_tf_available() else {} ) __A = True __A = False __A = False def UpperCamelCase_ (self ): """simple docstring""" a = TFBlenderbotModelTester(self ) a = ConfigTester(self , config_class=lowerCamelCase_ ) def UpperCamelCase_ (self ): """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase_ (self ): """simple docstring""" a = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowerCamelCase_ ) @require_tokenizers @require_tf class _lowercase ( unittest.TestCase ): """simple docstring""" __A = ["My friends are cool but they eat too many carbs."] __A = "facebook/blenderbot-400M-distill" @cached_property def UpperCamelCase_ (self ): """simple docstring""" return BlenderbotTokenizer.from_pretrained(self.model_name ) @cached_property def UpperCamelCase_ (self ): """simple docstring""" a = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model @slow def UpperCamelCase_ (self ): """simple docstring""" a = self.tokenizer(self.src_text , return_tensors="tf" ) a = self.model.generate( model_inputs.input_ids , ) a = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=lowerCamelCase_ )[0] assert ( generated_words == " That's unfortunate. Are they trying to lose weight or are they just trying to be healthier?" )
227
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase: str = logging.get_logger(__name__) _lowercase: List[str] = { "microsoft/trocr-base-handwritten": ( "https://huggingface.co/microsoft/trocr-base-handwritten/resolve/main/config.json" ), # See all TrOCR models at https://huggingface.co/models?filter=trocr } class _lowercase ( lowerCAmelCase ): """simple docstring""" __A = "trocr" __A = ["past_key_values"] __A = { "num_attention_heads": "decoder_attention_heads", "hidden_size": "d_model", "num_hidden_layers": "decoder_layers", } def __init__(self , lowerCamelCase_=50265 , lowerCamelCase_=1024 , lowerCamelCase_=12 , lowerCamelCase_=16 , lowerCamelCase_=4096 , lowerCamelCase_="gelu" , lowerCamelCase_=512 , lowerCamelCase_=0.1 , lowerCamelCase_=0.0 , lowerCamelCase_=0.0 , lowerCamelCase_=2 , lowerCamelCase_=0.02 , lowerCamelCase_=0.0 , lowerCamelCase_=True , lowerCamelCase_=False , lowerCamelCase_=True , lowerCamelCase_=True , lowerCamelCase_=1 , lowerCamelCase_=0 , lowerCamelCase_=2 , **lowerCamelCase_ , ): """simple docstring""" a = vocab_size a = d_model a = decoder_layers a = decoder_attention_heads a = decoder_ffn_dim a = activation_function a = max_position_embeddings a = dropout a = attention_dropout a = activation_dropout a = init_std a = decoder_layerdrop a = use_cache a = scale_embedding a = use_learned_position_embeddings a = layernorm_embedding super().__init__( pad_token_id=lowerCamelCase_ , bos_token_id=lowerCamelCase_ , eos_token_id=lowerCamelCase_ , decoder_start_token_id=lowerCamelCase_ , **lowerCamelCase_ , )
227
1
import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig a__ : Optional[Any] = logging.get_logger(__name__) class a_ : """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase ) ->Optional[int]: SCREAMING_SNAKE_CASE : Optional[Any] = question_encoder SCREAMING_SNAKE_CASE : str = generator SCREAMING_SNAKE_CASE : Any = self.question_encoder def __lowerCAmelCase ( self , _lowerCamelCase ) ->List[str]: if os.path.isfile(_lowerCamelCase ): raise ValueError(F"""Provided path ({save_directory}) should be a directory, not a file""" ) os.makedirs(_lowerCamelCase , exist_ok=_lowerCamelCase ) SCREAMING_SNAKE_CASE : Optional[int] = os.path.join(_lowerCamelCase , '''question_encoder_tokenizer''' ) SCREAMING_SNAKE_CASE : Dict = os.path.join(_lowerCamelCase , '''generator_tokenizer''' ) self.question_encoder.save_pretrained(_lowerCamelCase ) self.generator.save_pretrained(_lowerCamelCase ) @classmethod def __lowerCAmelCase ( cls , _lowerCamelCase , **_lowerCamelCase ) ->Dict: # dynamically import AutoTokenizer from ..auto.tokenization_auto import AutoTokenizer SCREAMING_SNAKE_CASE : Optional[Any] = kwargs.pop('''config''' , _lowerCamelCase ) if config is None: SCREAMING_SNAKE_CASE : Union[str, Any] = RagConfig.from_pretrained(_lowerCamelCase ) SCREAMING_SNAKE_CASE : Dict = AutoTokenizer.from_pretrained( _lowerCamelCase , config=config.question_encoder , subfolder='''question_encoder_tokenizer''' ) SCREAMING_SNAKE_CASE : int = AutoTokenizer.from_pretrained( _lowerCamelCase , config=config.generator , subfolder='''generator_tokenizer''' ) return cls(question_encoder=_lowerCamelCase , generator=_lowerCamelCase ) def __call__( self , *_lowerCamelCase , **_lowerCamelCase ) ->List[Any]: return self.current_tokenizer(*_lowerCamelCase , **_lowerCamelCase ) def __lowerCAmelCase ( self , *_lowerCamelCase , **_lowerCamelCase ) ->Union[str, Any]: return self.generator.batch_decode(*_lowerCamelCase , **_lowerCamelCase ) def __lowerCAmelCase ( self , *_lowerCamelCase , **_lowerCamelCase ) ->int: return self.generator.decode(*_lowerCamelCase , **_lowerCamelCase ) def __lowerCAmelCase ( self ) ->Optional[Any]: SCREAMING_SNAKE_CASE : List[str] = self.question_encoder def __lowerCAmelCase ( self ) ->Any: SCREAMING_SNAKE_CASE : Optional[Any] = self.generator def __lowerCAmelCase ( self , _lowerCamelCase , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = "longest" , _lowerCamelCase = None , _lowerCamelCase = True , **_lowerCamelCase , ) ->BatchEncoding: warnings.warn( '''`prepare_seq2seq_batch` is deprecated and will be removed in version 5 of 🤗 Transformers. Use the ''' '''regular `__call__` method to prepare your inputs and the tokenizer under the `with_target_tokenizer` ''' '''context manager to prepare your targets. See the documentation of your specific tokenizer for more ''' '''details''' , _lowerCamelCase , ) if max_length is None: SCREAMING_SNAKE_CASE : Union[str, Any] = self.current_tokenizer.model_max_length SCREAMING_SNAKE_CASE : Dict = self( _lowerCamelCase , add_special_tokens=_lowerCamelCase , return_tensors=_lowerCamelCase , max_length=_lowerCamelCase , padding=_lowerCamelCase , truncation=_lowerCamelCase , **_lowerCamelCase , ) if tgt_texts is None: return model_inputs # Process tgt_texts if max_target_length is None: SCREAMING_SNAKE_CASE : Any = self.current_tokenizer.model_max_length SCREAMING_SNAKE_CASE : Dict = self( text_target=_lowerCamelCase , add_special_tokens=_lowerCamelCase , return_tensors=_lowerCamelCase , padding=_lowerCamelCase , max_length=_lowerCamelCase , truncation=_lowerCamelCase , **_lowerCamelCase , ) SCREAMING_SNAKE_CASE : Optional[int] = labels['''input_ids'''] return model_inputs
19
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging a__ : Tuple = '''▁''' a__ : List[Any] = {'''vocab_file''': '''spiece.model'''} a__ : Optional[Any] = { '''vocab_file''': {'''google/pegasus-xsum''': '''https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model'''} } a__ : str = { '''google/pegasus-xsum''': 512, } a__ : str = logging.get_logger(__name__) class a_ ( a__ ): """simple docstring""" __SCREAMING_SNAKE_CASE : str = VOCAB_FILES_NAMES __SCREAMING_SNAKE_CASE : str = VOCAB_FILES_NAMES __SCREAMING_SNAKE_CASE : Tuple = PRETRAINED_VOCAB_FILES_MAP __SCREAMING_SNAKE_CASE : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __SCREAMING_SNAKE_CASE : str = ['input_ids', 'attention_mask'] def __init__( self , _lowerCamelCase , _lowerCamelCase="<pad>" , _lowerCamelCase="</s>" , _lowerCamelCase="<unk>" , _lowerCamelCase="<mask_2>" , _lowerCamelCase="<mask_1>" , _lowerCamelCase=None , _lowerCamelCase=103 , _lowerCamelCase = None , **_lowerCamelCase , ) ->None: SCREAMING_SNAKE_CASE : Dict = offset if additional_special_tokens is not None: if not isinstance(_lowerCamelCase , _lowerCamelCase ): raise TypeError( F"""additional_special_tokens should be of type {type(_lowerCamelCase )}, but is""" F""" {type(_lowerCamelCase )}""" ) SCREAMING_SNAKE_CASE : List[Any] = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ F"""<unk_{i}>""" for i in range(len(_lowerCamelCase ) , self.offset - 1 ) ] if len(set(_lowerCamelCase ) ) != len(_lowerCamelCase ): raise ValueError( '''Please make sure that the provided additional_special_tokens do not contain an incorrectly''' F""" shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.""" ) SCREAMING_SNAKE_CASE : Dict = additional_special_tokens_extended else: SCREAMING_SNAKE_CASE : str = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [F"""<unk_{i}>""" for i in range(2 , self.offset )] SCREAMING_SNAKE_CASE : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=_lowerCamelCase , unk_token=_lowerCamelCase , mask_token=_lowerCamelCase , pad_token=_lowerCamelCase , mask_token_sent=_lowerCamelCase , offset=_lowerCamelCase , additional_special_tokens=_lowerCamelCase , sp_model_kwargs=self.sp_model_kwargs , **_lowerCamelCase , ) SCREAMING_SNAKE_CASE : List[str] = mask_token_sent SCREAMING_SNAKE_CASE : Optional[int] = vocab_file SCREAMING_SNAKE_CASE : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(_lowerCamelCase ) # add special tokens to encoder dict SCREAMING_SNAKE_CASE : Dict[int, str] = { 0: self.pad_token, 1: self.eos_token, } if self.mask_token_sent is not None: self.encoder.update( { 2: self.mask_token_sent, 3: self.mask_token, } ) if self.offset > 0: # entries 2-104 are only used for pretraining and called <mask_1>, <mask_2>, unk_2, ...unk_102 # mask_token_sent is already added to list -> so start at 1 self.encoder.update({i + 3: additional_special_tokens[i] for i in range(1 , self.offset - 1 )} ) SCREAMING_SNAKE_CASE : Dict[str, int] = {v: k for k, v in self.encoder.items()} @property def __lowerCAmelCase ( self ) ->int: return len(self.sp_model ) + self.offset def __lowerCAmelCase ( self ) ->Dict[str, int]: SCREAMING_SNAKE_CASE : Union[str, Any] = {self.convert_ids_to_tokens(_lowerCamelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ) ->Optional[Any]: SCREAMING_SNAKE_CASE : Optional[int] = self.__dict__.copy() SCREAMING_SNAKE_CASE : str = None return state def __setstate__( self , _lowerCamelCase ) ->Union[str, Any]: SCREAMING_SNAKE_CASE : Optional[Any] = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): SCREAMING_SNAKE_CASE : List[str] = {} SCREAMING_SNAKE_CASE : str = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def __lowerCAmelCase ( self , _lowerCamelCase ) ->List[str]: return self.sp_model.encode(_lowerCamelCase , out_type=_lowerCamelCase ) def __lowerCAmelCase ( self , _lowerCamelCase ) ->int: if token in self.decoder: return self.decoder[token] elif token in self.added_tokens_decoder: return self.added_tokens_decoder[token] SCREAMING_SNAKE_CASE : List[str] = self.sp_model.piece_to_id(_lowerCamelCase ) return sp_id + self.offset def __lowerCAmelCase ( self , _lowerCamelCase ) ->str: if index in self.encoder: return self.encoder[index] elif index in self.added_tokens_encoder: return self.added_tokens_encoder[index] else: SCREAMING_SNAKE_CASE : Dict = self.sp_model.IdToPiece(index - self.offset ) return token def __lowerCAmelCase ( self , _lowerCamelCase ) ->Union[str, Any]: SCREAMING_SNAKE_CASE : Dict = [] SCREAMING_SNAKE_CASE : int = '''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(_lowerCamelCase ) + token SCREAMING_SNAKE_CASE : Optional[Any] = [] else: current_sub_tokens.append(_lowerCamelCase ) out_string += self.sp_model.decode(_lowerCamelCase ) return out_string.strip() def __lowerCAmelCase ( self , _lowerCamelCase=False ) ->str: return 1 def __lowerCAmelCase ( self , _lowerCamelCase ) ->int: SCREAMING_SNAKE_CASE : Dict = set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special return [1 if x in all_special_ids else 0 for x in seq] def __lowerCAmelCase ( self , _lowerCamelCase , _lowerCamelCase = None , _lowerCamelCase = False ) ->List[int]: if already_has_special_tokens: return self._special_token_mask(_lowerCamelCase ) elif token_ids_a is None: return self._special_token_mask(_lowerCamelCase ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def __lowerCAmelCase ( self , _lowerCamelCase , _lowerCamelCase=None ) ->List[int]: if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def __lowerCAmelCase ( self , _lowerCamelCase , _lowerCamelCase = None ) ->Tuple[str]: if not os.path.isdir(_lowerCamelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return SCREAMING_SNAKE_CASE : int = os.path.join( _lowerCamelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_lowerCamelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _lowerCamelCase ) elif not os.path.isfile(self.vocab_file ): with open(_lowerCamelCase , '''wb''' ) as fi: SCREAMING_SNAKE_CASE : Tuple = self.sp_model.serialized_model_proto() fi.write(_lowerCamelCase ) return (out_vocab_file,)
19
1
"""simple docstring""" from manim import * class lowerCAmelCase_ ( A__ ): '''simple docstring''' def A__ ( self ) -> Optional[int]: __lowerCAmelCase = Rectangle(height=0.5 , width=0.5 ) __lowerCAmelCase = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) __lowerCAmelCase = [mem.copy() for i in range(6 )] __lowerCAmelCase = [mem.copy() for i in range(6 )] __lowerCAmelCase = VGroup(*snake_case_ ).arrange(snake_case_ , buff=0 ) __lowerCAmelCase = VGroup(*snake_case_ ).arrange(snake_case_ , buff=0 ) __lowerCAmelCase = VGroup(snake_case_ , snake_case_ ).arrange(snake_case_ , buff=0 ) __lowerCAmelCase = Text("""CPU""" , font_size=24 ) __lowerCAmelCase = Group(snake_case_ , snake_case_ ).arrange(snake_case_ , buff=0.5 , aligned_edge=snake_case_ ) cpu.move_to([-2.5, -0.5, 0] ) self.add(snake_case_ ) __lowerCAmelCase = [mem.copy() for i in range(1 )] __lowerCAmelCase = VGroup(*snake_case_ ).arrange(snake_case_ , buff=0 ) __lowerCAmelCase = Text("""GPU""" , font_size=24 ) __lowerCAmelCase = Group(snake_case_ , snake_case_ ).arrange(snake_case_ , buff=0.5 , aligned_edge=snake_case_ ) gpu.align_to(snake_case_ , snake_case_ ) gpu.set_x(gpu.get_x() - 1 ) self.add(snake_case_ ) __lowerCAmelCase = [mem.copy() for i in range(6 )] __lowerCAmelCase = VGroup(*snake_case_ ).arrange(snake_case_ , buff=0 ) __lowerCAmelCase = Text("""Model""" , font_size=24 ) __lowerCAmelCase = Group(snake_case_ , snake_case_ ).arrange(snake_case_ , buff=0.5 , aligned_edge=snake_case_ ) model.move_to([3, -1.0, 0] ) self.play( Create(snake_case_ , run_time=1 ) , Create(snake_case_ , run_time=1 ) , Create(snake_case_ , run_time=1 ) , ) __lowerCAmelCase = MarkupText( f"""First, an empty model skeleton is loaded\ninto <span fgcolor='{YELLOW}'>memory</span> without using much RAM.""" , font_size=24 , ) __lowerCAmelCase = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) __lowerCAmelCase = MarkupText( f"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) step_a.move_to([2, 2, 0] ) self.play(Write(snake_case_ , run_time=2.5 ) , Write(snake_case_ ) , Write(snake_case_ ) ) self.add(snake_case_ ) __lowerCAmelCase = [] __lowerCAmelCase = [] __lowerCAmelCase = [] for i, rect in enumerate(snake_case_ ): __lowerCAmelCase = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0.0 ).set_fill(snake_case_ , opacity=0.7 ) cpu_target.move_to(snake_case_ ) cpu_target.generate_target() __lowerCAmelCase = 0.46 / 4 __lowerCAmelCase = 0.46 / 3 if i == 0: cpu_target.target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=snake_case_ ) cpu_target.target.set_x(cpu_target.target.get_x() + 0.1 ) elif i == 3: cpu_target.target.next_to(cpu_targs[0].target , direction=snake_case_ , buff=0.0 ) else: cpu_target.target.next_to(cpu_targs[i - 1].target , direction=snake_case_ , buff=0.0 ) cpu_targs.append(snake_case_ ) first_animations.append(rect.animate(run_time=0.5 ).set_stroke(snake_case_ ) ) second_animations.append(MoveToTarget(snake_case_ , run_time=1.5 ) ) self.play(*snake_case_ ) self.play(*snake_case_ ) self.wait()
301
"""simple docstring""" def lowercase (_lowerCAmelCase = 100_0000 ): __lowerCAmelCase = 1 __lowerCAmelCase = 1 __lowerCAmelCase = {1: 1} for inputa in range(2 , _lowerCAmelCase ): __lowerCAmelCase = 0 __lowerCAmelCase = inputa while True: if number in counters: counter += counters[number] break if number % 2 == 0: number //= 2 counter += 1 else: __lowerCAmelCase = (3 * number) + 1 counter += 1 if inputa not in counters: __lowerCAmelCase = counter if counter > pre_counter: __lowerCAmelCase = inputa __lowerCAmelCase = counter return largest_number if __name__ == "__main__": print(solution(int(input().strip())))
301
1
"""simple docstring""" def lowercase ( _SCREAMING_SNAKE_CASE : list ): '''simple docstring''' for i in range(len(_SCREAMING_SNAKE_CASE ) - 1 , 0 , -1 ): _UpperCAmelCase = False for j in range(_SCREAMING_SNAKE_CASE , 0 , -1 ): if unsorted[j] < unsorted[j - 1]: _UpperCAmelCase , _UpperCAmelCase = unsorted[j - 1], unsorted[j] _UpperCAmelCase = True for j in range(_SCREAMING_SNAKE_CASE ): if unsorted[j] > unsorted[j + 1]: _UpperCAmelCase , _UpperCAmelCase = unsorted[j + 1], unsorted[j] _UpperCAmelCase = True if not swapped: break return unsorted if __name__ == "__main__": import doctest doctest.testmod() __A : Optional[int] = input("Enter numbers separated by a comma:\n").strip() __A : str = [int(item) for item in user_input.split(",")] print(f'''{cocktail_shaker_sort(unsorted) = }''')
326
"""simple docstring""" import json import os import unittest from transformers.models.ctrl.tokenization_ctrl import VOCAB_FILES_NAMES, CTRLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class _a ( lowerCAmelCase , unittest.TestCase): """simple docstring""" UpperCamelCase__ = CTRLTokenizer UpperCamelCase__ = False UpperCamelCase__ = False def lowercase__ ( self : Dict )->str: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt _UpperCAmelCase = ['''adapt''', '''re@@''', '''a@@''', '''apt''', '''c@@''', '''t''', '''<unk>'''] _UpperCAmelCase = dict(zip(__UpperCamelCase , range(len(__UpperCamelCase ) ) ) ) _UpperCAmelCase = ['''#version: 0.2''', '''a p''', '''ap t</w>''', '''r e''', '''a d''', '''ad apt</w>''', ''''''] _UpperCAmelCase = {'''unk_token''': '''<unk>'''} _UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) _UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(__UpperCamelCase ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(__UpperCamelCase ) ) def lowercase__ ( self : str , **__UpperCamelCase : Union[str, Any] )->Any: kwargs.update(self.special_tokens_map ) return CTRLTokenizer.from_pretrained(self.tmpdirname , **__UpperCamelCase ) def lowercase__ ( self : Union[str, Any] , __UpperCamelCase : Optional[int] )->Tuple: _UpperCAmelCase = '''adapt react readapt apt''' _UpperCAmelCase = '''adapt react readapt apt''' return input_text, output_text def lowercase__ ( self : Dict )->Optional[int]: _UpperCAmelCase = CTRLTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) _UpperCAmelCase = '''adapt react readapt apt''' _UpperCAmelCase = '''adapt re@@ a@@ c@@ t re@@ adapt apt'''.split() _UpperCAmelCase = tokenizer.tokenize(__UpperCamelCase ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) _UpperCAmelCase = tokens + [tokenizer.unk_token] _UpperCAmelCase = [0, 1, 2, 4, 5, 1, 0, 3, 6] self.assertListEqual(tokenizer.convert_tokens_to_ids(__UpperCamelCase ) , __UpperCamelCase )
326
1
import unittest from transformers import BigBirdTokenizer, BigBirdTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin snake_case : Any = '''▁''' snake_case : Optional[Any] = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece @require_tokenizers class _snake_case ( _snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE__ = BigBirdTokenizer SCREAMING_SNAKE_CASE__ = BigBirdTokenizerFast SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = True def SCREAMING_SNAKE_CASE__ ( self ): super().setUp() a :Union[str, Any] = self.tokenizer_class(_lowerCamelCase , keep_accents=_lowerCamelCase ) tokenizer.save_pretrained(self.tmpdirname ) def SCREAMING_SNAKE_CASE__ ( self ): a :Dict = '''<s>''' a :int = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_lowerCamelCase ) , _lowerCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_lowerCamelCase ) , _lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ): a :int = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<unk>''' ) self.assertEqual(vocab_keys[1] , '''<s>''' ) self.assertEqual(vocab_keys[-1] , '''[MASK]''' ) self.assertEqual(len(_lowerCamelCase ) , 1004 ) def SCREAMING_SNAKE_CASE__ ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1000 ) def SCREAMING_SNAKE_CASE__ ( self ): if not self.test_rust_tokenizer: return a :Union[str, Any] = self.get_tokenizer() a :Dict = self.get_rust_tokenizer() a :Union[str, Any] = '''I was born in 92000, and this is falsé.''' a :List[str] = tokenizer.tokenize(_lowerCamelCase ) a :Union[str, Any] = rust_tokenizer.tokenize(_lowerCamelCase ) self.assertListEqual(_lowerCamelCase , _lowerCamelCase ) a :List[Any] = tokenizer.encode(_lowerCamelCase , add_special_tokens=_lowerCamelCase ) a :str = rust_tokenizer.encode(_lowerCamelCase , add_special_tokens=_lowerCamelCase ) self.assertListEqual(_lowerCamelCase , _lowerCamelCase ) a :str = self.get_rust_tokenizer() a :Optional[Any] = tokenizer.encode(_lowerCamelCase ) a :Dict = rust_tokenizer.encode(_lowerCamelCase ) self.assertListEqual(_lowerCamelCase , _lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ): a :str = BigBirdTokenizer(_lowerCamelCase , keep_accents=_lowerCamelCase ) a :str = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(_lowerCamelCase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_lowerCamelCase ) , [285, 46, 10, 170, 382] , ) a :List[Any] = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( _lowerCamelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) a :List[Any] = tokenizer.convert_tokens_to_ids(_lowerCamelCase ) self.assertListEqual( _lowerCamelCase , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) a :Union[str, Any] = tokenizer.convert_ids_to_tokens(_lowerCamelCase ) self.assertListEqual( _lowerCamelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) @cached_property def SCREAMING_SNAKE_CASE__ ( self ): return BigBirdTokenizer.from_pretrained('''google/bigbird-roberta-base''' ) @slow def SCREAMING_SNAKE_CASE__ ( self ): a :int = '''Hello World!''' a :Optional[int] = [65, 1_8536, 2260, 101, 66] self.assertListEqual(_lowerCamelCase , self.big_tokenizer.encode(_lowerCamelCase ) ) @slow def SCREAMING_SNAKE_CASE__ ( self ): a :Union[str, Any] = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) # fmt: off a :str = [65, 871, 419, 358, 946, 991, 2521, 452, 358, 1357, 387, 7751, 3536, 112, 985, 456, 126, 865, 938, 5400, 5734, 458, 1368, 467, 786, 2462, 5246, 1159, 633, 865, 4519, 457, 582, 852, 2557, 427, 916, 508, 405, 3_4324, 497, 391, 408, 1_1342, 1244, 385, 100, 938, 985, 456, 574, 362, 1_2597, 3200, 3129, 1172, 66] # noqa: E231 # fmt: on self.assertListEqual(_lowerCamelCase , self.big_tokenizer.encode(_lowerCamelCase ) ) @require_torch @slow def SCREAMING_SNAKE_CASE__ ( self ): import torch from transformers import BigBirdConfig, BigBirdModel # Build sequence a :Union[str, Any] = list(self.big_tokenizer.get_vocab().keys() )[:10] a :Dict = ''' '''.join(_lowerCamelCase ) a :Tuple = self.big_tokenizer.encode_plus(_lowerCamelCase , return_tensors='''pt''' , return_token_type_ids=_lowerCamelCase ) a :int = self.big_tokenizer.batch_encode_plus( [sequence + ''' ''' + sequence] , return_tensors='''pt''' , return_token_type_ids=_lowerCamelCase ) a :Union[str, Any] = BigBirdConfig(attention_type='''original_full''' ) a :List[str] = BigBirdModel(_lowerCamelCase ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**_lowerCamelCase ) model(**_lowerCamelCase ) @slow def SCREAMING_SNAKE_CASE__ ( self ): a :int = BigBirdTokenizer.from_pretrained('''google/bigbird-roberta-base''' ) a :str = tokenizer.decode(tokenizer('''Paris is the [MASK].''' ).input_ids ) self.assertTrue(decoded_text == '''[CLS] Paris is the[MASK].[SEP]''' ) @slow def SCREAMING_SNAKE_CASE__ ( self ): # fmt: off a :List[Any] = {'''input_ids''': [[65, 3_9286, 458, 3_6335, 2001, 456, 1_3073, 1_3266, 455, 113, 7746, 1741, 1_1157, 391, 1_3073, 1_3266, 455, 113, 3967, 3_5412, 113, 4936, 109, 3870, 2377, 113, 3_0084, 4_5720, 458, 134, 1_7496, 112, 503, 1_1672, 113, 118, 112, 5665, 1_3347, 3_8687, 112, 1496, 3_1389, 112, 3268, 4_7264, 134, 962, 112, 1_6377, 8035, 2_3130, 430, 1_2169, 1_5518, 2_8592, 458, 146, 4_1697, 109, 391, 1_2169, 1_5518, 1_6689, 458, 146, 4_1358, 109, 452, 726, 4034, 111, 763, 3_5412, 5082, 388, 1903, 111, 9051, 391, 2870, 4_8918, 1900, 1123, 550, 998, 112, 9586, 1_5985, 455, 391, 410, 2_2955, 3_7636, 114, 66], [65, 448, 1_7496, 419, 3663, 385, 763, 113, 2_7533, 2870, 3283, 1_3043, 1639, 2_4713, 523, 656, 2_4013, 1_8550, 2521, 517, 2_7014, 2_1244, 420, 1212, 1465, 391, 927, 4833, 388, 578, 1_1786, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [65, 484, 2169, 7687, 2_1932, 1_8146, 726, 363, 1_7032, 3391, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_lowerCamelCase , model_name='''google/bigbird-roberta-base''' , revision='''215c99f1600e06f83acce68422f2035b2b5c3510''' , )
94
import string import numpy def __lowerCamelCase ( UpperCAmelCase_ : int , UpperCAmelCase_ : int ): """simple docstring""" return b if a == 0 else greatest_common_divisor(b % a , UpperCAmelCase_ ) class _snake_case : SCREAMING_SNAKE_CASE__ = string.ascii_uppercase + string.digits # This cipher takes alphanumerics into account # i.e. a total of 36 characters # take x and return x % len(key_string) SCREAMING_SNAKE_CASE__ = numpy.vectorize(lambda _snake_case : x % 36 ) SCREAMING_SNAKE_CASE__ = numpy.vectorize(_snake_case ) def __init__( self , _lowerCamelCase ): a :List[Any] = self.modulus(_lowerCamelCase ) # mod36 calc's on the encrypt key self.check_determinant() # validate the determinant of the encryption key a :int = encrypt_key.shape[0] def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase ): return self.key_string.index(_lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase ): return self.key_string[round(_lowerCamelCase )] def SCREAMING_SNAKE_CASE__ ( self ): a :str = round(numpy.linalg.det(self.encrypt_key ) ) if det < 0: a :Any = det % len(self.key_string ) a :Dict = len(self.key_string ) if greatest_common_divisor(_lowerCamelCase , len(self.key_string ) ) != 1: a :int = ( F'''determinant modular {req_l} of encryption key({det}) ''' F'''is not co prime w.r.t {req_l}.\nTry another key.''' ) raise ValueError(_lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase ): a :Optional[Any] = [char for char in text.upper() if char in self.key_string] a :List[str] = chars[-1] while len(_lowerCamelCase ) % self.break_key != 0: chars.append(_lowerCamelCase ) return "".join(_lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase ): a :Dict = self.process_text(text.upper() ) a :List[str] = '''''' for i in range(0 , len(_lowerCamelCase ) - self.break_key + 1 , self.break_key ): a :int = text[i : i + self.break_key] a :Optional[int] = [self.replace_letters(_lowerCamelCase ) for char in batch] a :Union[str, Any] = numpy.array([vec] ).T a :str = self.modulus(self.encrypt_key.dot(_lowerCamelCase ) ).T.tolist()[ 0 ] a :List[Any] = ''''''.join( self.replace_digits(_lowerCamelCase ) for num in batch_encrypted ) encrypted += encrypted_batch return encrypted def SCREAMING_SNAKE_CASE__ ( self ): a :List[Any] = round(numpy.linalg.det(self.encrypt_key ) ) if det < 0: a :int = det % len(self.key_string ) a :Tuple = None for i in range(len(self.key_string ) ): if (det * i) % len(self.key_string ) == 1: a :Tuple = i break a :List[str] = ( det_inv * numpy.linalg.det(self.encrypt_key ) * numpy.linalg.inv(self.encrypt_key ) ) return self.to_int(self.modulus(_lowerCamelCase ) ) def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase ): a :List[Any] = self.make_decrypt_key() a :str = self.process_text(text.upper() ) a :List[Any] = '''''' for i in range(0 , len(_lowerCamelCase ) - self.break_key + 1 , self.break_key ): a :Optional[Any] = text[i : i + self.break_key] a :List[Any] = [self.replace_letters(_lowerCamelCase ) for char in batch] a :str = numpy.array([vec] ).T a :Dict = self.modulus(decrypt_key.dot(_lowerCamelCase ) ).T.tolist()[0] a :List[Any] = ''''''.join( self.replace_digits(_lowerCamelCase ) for num in batch_decrypted ) decrypted += decrypted_batch return decrypted def __lowerCamelCase ( ): """simple docstring""" a :Tuple = int(input('''Enter the order of the encryption key: ''' ) ) a :Dict = [] print('''Enter each row of the encryption key with space separated integers''' ) for _ in range(UpperCAmelCase_ ): a :List[str] = [int(UpperCAmelCase_ ) for x in input().split()] hill_matrix.append(UpperCAmelCase_ ) a :Any = HillCipher(numpy.array(UpperCAmelCase_ ) ) print('''Would you like to encrypt or decrypt some text? (1 or 2)''' ) a :Any = input('''\n1. Encrypt\n2. Decrypt\n''' ) if option == "1": a :str = input('''What text would you like to encrypt?: ''' ) print('''Your encrypted text is:''' ) print(hc.encrypt(UpperCAmelCase_ ) ) elif option == "2": a :Dict = input('''What text would you like to decrypt?: ''' ) print('''Your decrypted text is:''' ) print(hc.decrypt(UpperCAmelCase_ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
94
1
'''simple docstring''' __UpperCAmelCase = { '''Pillow''': '''Pillow''', '''accelerate''': '''accelerate>=0.11.0''', '''compel''': '''compel==0.1.8''', '''black''': '''black~=23.1''', '''datasets''': '''datasets''', '''filelock''': '''filelock''', '''flax''': '''flax>=0.4.1''', '''hf-doc-builder''': '''hf-doc-builder>=0.3.0''', '''huggingface-hub''': '''huggingface-hub>=0.13.2''', '''requests-mock''': '''requests-mock==1.10.0''', '''importlib_metadata''': '''importlib_metadata''', '''invisible-watermark''': '''invisible-watermark''', '''isort''': '''isort>=5.5.4''', '''jax''': '''jax>=0.2.8,!=0.3.2''', '''jaxlib''': '''jaxlib>=0.1.65''', '''Jinja2''': '''Jinja2''', '''k-diffusion''': '''k-diffusion>=0.0.12''', '''torchsde''': '''torchsde''', '''note_seq''': '''note_seq''', '''librosa''': '''librosa''', '''numpy''': '''numpy''', '''omegaconf''': '''omegaconf''', '''parameterized''': '''parameterized''', '''protobuf''': '''protobuf>=3.20.3,<4''', '''pytest''': '''pytest''', '''pytest-timeout''': '''pytest-timeout''', '''pytest-xdist''': '''pytest-xdist''', '''ruff''': '''ruff>=0.0.241''', '''safetensors''': '''safetensors''', '''sentencepiece''': '''sentencepiece>=0.1.91,!=0.1.92''', '''scipy''': '''scipy''', '''onnx''': '''onnx''', '''regex''': '''regex!=2019.12.17''', '''requests''': '''requests''', '''tensorboard''': '''tensorboard''', '''torch''': '''torch>=1.4''', '''torchvision''': '''torchvision''', '''transformers''': '''transformers>=4.25.1''', '''urllib3''': '''urllib3<=2.0.0''', }
365
'''simple docstring''' import argparse import json import subprocess def _snake_case ( A , A ) -> Tuple: lowerCAmelCase__ = [] lowerCAmelCase__ = ( F"""curl -H \"Accept: application/vnd.github+json\" -H \"Authorization: Bearer {token}\"""" ''' https://api.github.com/repos/huggingface/transformers/actions/runners''' ) lowerCAmelCase__ = subprocess.run(A , shell=A , stdout=subprocess.PIPE ) lowerCAmelCase__ = output.stdout.decode('''utf-8''' ) lowerCAmelCase__ = json.loads(A ) lowerCAmelCase__ = status['''runners'''] for runner in runners: if runner["name"] in target_runners: if runner["status"] == "offline": offline_runners.append(A ) # save the result so we can report them on Slack with open('''offline_runners.txt''' , '''w''' ) as fp: fp.write(json.dumps(A ) ) if len(A ) > 0: lowerCAmelCase__ = '''\n'''.join([x['''name'''] for x in offline_runners] ) raise ValueError(F"""The following runners are offline:\n{failed}""" ) if __name__ == "__main__": def _snake_case ( A ) -> Optional[Any]: return values.split(''',''' ) __UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--target_runners''', default=None, type=list_str, required=True, help='''Comma-separated list of runners to check status.''', ) parser.add_argument( '''--token''', default=None, type=str, required=True, help='''A token that has actions:read permission.''' ) __UpperCAmelCase = parser.parse_args() get_runner_status(args.target_runners, args.token)
228
0
"""simple docstring""" import warnings from diffusers import StableDiffusionInpaintPipeline as StableDiffusionInpaintPipeline # noqa F401 warnings.warn( """The `inpainting.py` script is outdated. Please use directly `from diffusers import""" """ StableDiffusionInpaintPipeline` instead.""" )
102
from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class SCREAMING_SNAKE_CASE_ : def __init__( self : Optional[int] , lowerCamelCase_ : Optional[Any] , lowerCamelCase_ : str=3 , lowerCamelCase_ : Tuple=32 , lowerCamelCase_ : List[str]=3 , lowerCamelCase_ : Optional[int]=10 , lowerCamelCase_ : List[str]=[10, 20, 30, 40] , lowerCamelCase_ : Tuple=[1, 1, 2, 1] , lowerCamelCase_ : Dict=True , lowerCamelCase_ : str=True , lowerCamelCase_ : Tuple="relu" , lowerCamelCase_ : List[str]=3 , lowerCamelCase_ : Dict=None , ): """simple docstring""" UpperCamelCase = parent UpperCamelCase = batch_size UpperCamelCase = image_size UpperCamelCase = num_channels UpperCamelCase = embeddings_size UpperCamelCase = hidden_sizes UpperCamelCase = depths UpperCamelCase = is_training UpperCamelCase = use_labels UpperCamelCase = hidden_act UpperCamelCase = num_labels UpperCamelCase = scope UpperCamelCase = len(lowerCamelCase_ ) def lowerCamelCase_ ( self : Dict ): """simple docstring""" UpperCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase = None if self.use_labels: UpperCamelCase = ids_tensor([self.batch_size] , self.num_labels ) UpperCamelCase = self.get_config() return config, pixel_values, labels def lowerCamelCase_ ( self : Optional[int] ): """simple docstring""" return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def lowerCamelCase_ ( self : Any , lowerCamelCase_ : List[Any] , lowerCamelCase_ : Optional[int] , lowerCamelCase_ : Tuple ): """simple docstring""" UpperCamelCase = TFResNetModel(config=lowerCamelCase_ ) UpperCamelCase = model(lowerCamelCase_ ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def lowerCamelCase_ ( self : List[Any] , lowerCamelCase_ : List[str] , lowerCamelCase_ : str , lowerCamelCase_ : Optional[int] ): """simple docstring""" UpperCamelCase = self.num_labels UpperCamelCase = TFResNetForImageClassification(lowerCamelCase_ ) UpperCamelCase = model(lowerCamelCase_ , labels=lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase_ ( self : int ): """simple docstring""" UpperCamelCase = self.prepare_config_and_inputs() UpperCamelCase , UpperCamelCase , UpperCamelCase = config_and_inputs UpperCamelCase = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class SCREAMING_SNAKE_CASE_ ( __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): __lowerCAmelCase = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () __lowerCAmelCase = ( {"""feature-extraction""": TFResNetModel, """image-classification""": TFResNetForImageClassification} if is_tf_available() else {} ) __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False __lowerCAmelCase = False def lowerCamelCase_ ( self : Any ): """simple docstring""" UpperCamelCase = TFResNetModelTester(self ) UpperCamelCase = ConfigTester(self , config_class=lowerCamelCase_ , has_text_modality=lowerCamelCase_ ) def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowerCamelCase_ ( self : Union[str, Any] ): """simple docstring""" return @unittest.skip(reason="""ResNet does not use inputs_embeds""" ) def lowerCamelCase_ ( self : Dict ): """simple docstring""" pass @unittest.skip(reason="""ResNet does not support input and output embeddings""" ) def lowerCamelCase_ ( self : Optional[Any] ): """simple docstring""" pass def lowerCamelCase_ ( self : Optional[int] ): """simple docstring""" UpperCamelCase , UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase = model_class(lowerCamelCase_ ) UpperCamelCase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase = [*signature.parameters.keys()] UpperCamelCase = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , lowerCamelCase_ ) def lowerCamelCase_ ( self : List[str] ): """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase_ ) def lowerCamelCase_ ( self : Union[str, Any] ): """simple docstring""" def check_hidden_states_output(lowerCamelCase_ : Tuple , lowerCamelCase_ : int , lowerCamelCase_ : str ): UpperCamelCase = model_class(lowerCamelCase_ ) UpperCamelCase = model(**self._prepare_for_class(lowerCamelCase_ , lowerCamelCase_ ) ) UpperCamelCase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states UpperCamelCase = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase_ ) , expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) UpperCamelCase , UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase = ["""basic""", """bottleneck"""] for model_class in self.all_model_classes: for layer_type in layers_type: UpperCamelCase = layer_type UpperCamelCase = True check_hidden_states_output(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase = True check_hidden_states_output(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase_ ) @slow def lowerCamelCase_ ( self : Any ): """simple docstring""" for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase = TFResNetModel.from_pretrained(lowerCamelCase_ ) self.assertIsNotNone(lowerCamelCase_ ) def lowercase( ) -> Any: '''simple docstring''' UpperCamelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): @cached_property def lowerCamelCase_ ( self : Dict ): """simple docstring""" return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def lowerCamelCase_ ( self : Dict ): """simple docstring""" UpperCamelCase = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) UpperCamelCase = self.default_image_processor UpperCamelCase = prepare_img() UpperCamelCase = image_processor(images=lowerCamelCase_ , return_tensors="""tf""" ) # forward pass UpperCamelCase = model(**lowerCamelCase_ ) # verify the logits UpperCamelCase = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase_ ) UpperCamelCase = tf.constant([-1_1.1_0_6_9, -9.7_8_7_7, -8.3_7_7_7] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , lowerCamelCase_ , atol=1E-4 ) )
343
0
import inspect import os import unittest from dataclasses import dataclass import torch from accelerate import Accelerator, DistributedDataParallelKwargs, GradScalerKwargs from accelerate.state import AcceleratorState from accelerate.test_utils import execute_subprocess_async, require_cuda, require_multi_gpu from accelerate.utils import KwargsHandler @dataclass class __lowercase (UpperCamelCase__ ): """simple docstring""" _snake_case = 0 _snake_case = False _snake_case = 3.0 class __lowercase (unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> Dict: # If no defaults are changed, `to_kwargs` returns an empty dict. self.assertDictEqual(MockClass().to_kwargs() , {} ) self.assertDictEqual(MockClass(a=2 ).to_kwargs() , {"""a""": 2} ) self.assertDictEqual(MockClass(a=2 , b=A ).to_kwargs() , {"""a""": 2, """b""": True} ) self.assertDictEqual(MockClass(a=2 , c=2.25 ).to_kwargs() , {"""a""": 2, """c""": 2.25} ) @require_cuda def UpperCAmelCase ( self ) -> Optional[int]: # If no defaults are changed, `to_kwargs` returns an empty dict. snake_case : List[Any] = GradScalerKwargs(init_scale=1_0_2_4 , growth_factor=2 ) AcceleratorState._reset_state() snake_case : Optional[int] = Accelerator(mixed_precision="""fp16""" , kwargs_handlers=[scaler_handler] ) print(accelerator.use_fpaa ) snake_case : Optional[Any] = accelerator.scaler # Check the kwargs have been applied self.assertEqual(scaler._init_scale , 10_24.0 ) self.assertEqual(scaler._growth_factor , 2.0 ) # Check the other values are at the default self.assertEqual(scaler._backoff_factor , 0.5 ) self.assertEqual(scaler._growth_interval , 2_0_0_0 ) self.assertEqual(scaler._enabled , A ) @require_multi_gpu def UpperCAmelCase ( self ) -> Tuple: snake_case : Dict = ["""torchrun""", f"""--nproc_per_node={torch.cuda.device_count()}""", inspect.getfile(self.__class__ )] execute_subprocess_async(A , env=os.environ.copy() ) if __name__ == "__main__": lowerCamelCase : List[str] = DistributedDataParallelKwargs(bucket_cap_mb=1_5, find_unused_parameters=True) lowerCamelCase : Dict = Accelerator(kwargs_handlers=[ddp_scaler]) lowerCamelCase : List[str] = torch.nn.Linear(1_0_0, 2_0_0) lowerCamelCase : str = accelerator.prepare(model) # Check the values changed in kwargs lowerCamelCase : Union[str, Any] = '' lowerCamelCase : Dict = model.bucket_bytes_cap // (1_0_2_4 * 1_0_2_4) if observed_bucket_cap_map != 1_5: error_msg += f"Kwargs badly passed, should have `15` but found {observed_bucket_cap_map}.\n" if model.find_unused_parameters is not True: error_msg += f"Kwargs badly passed, should have `True` but found {model.find_unused_parameters}.\n" # Check the values of the defaults if model.dim != 0: error_msg += f"Default value not respected, should have `0` but found {model.dim}.\n" if model.broadcast_buffers is not True: error_msg += f"Default value not respected, should have `True` but found {model.broadcast_buffers}.\n" if model.gradient_as_bucket_view is not False: error_msg += f"Default value not respected, should have `False` but found {model.gradient_as_bucket_view}.\n" # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
176
import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class __lowercase (UpperCamelCase__ , unittest.TestCase ): """simple docstring""" _snake_case = ConsistencyModelPipeline _snake_case = UNCONDITIONAL_IMAGE_GENERATION_PARAMS _snake_case = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt _snake_case = frozenset( [ """num_inference_steps""", """generator""", """latents""", """output_type""", """return_dict""", """callback""", """callback_steps""", ] ) @property def UpperCAmelCase ( self ) -> List[str]: snake_case : Dict = UNetaDModel.from_pretrained( """diffusers/consistency-models-test""" , subfolder="""test_unet""" , ) return unet @property def UpperCAmelCase ( self ) -> Any: snake_case : Optional[int] = UNetaDModel.from_pretrained( """diffusers/consistency-models-test""" , subfolder="""test_unet_class_cond""" , ) return unet def UpperCAmelCase ( self , A=False ) -> Optional[Any]: if class_cond: snake_case : List[str] = self.dummy_cond_unet else: snake_case : Optional[Any] = self.dummy_uncond_unet # Default to CM multistep sampler snake_case : Optional[Any] = CMStochasticIterativeScheduler( num_train_timesteps=4_0 , sigma_min=0.0_02 , sigma_max=80.0 , ) snake_case : Dict = { """unet""": unet, """scheduler""": scheduler, } return components def UpperCAmelCase ( self , A , A=0 ) -> Optional[int]: if str(A ).startswith("""mps""" ): snake_case : Union[str, Any] = torch.manual_seed(A ) else: snake_case : Dict = torch.Generator(device=A ).manual_seed(A ) snake_case : Tuple = { """batch_size""": 1, """num_inference_steps""": None, """timesteps""": [2_2, 0], """generator""": generator, """output_type""": """np""", } return inputs def UpperCAmelCase ( self ) -> List[Any]: snake_case : Union[str, Any] = """cpu""" # ensure determinism for the device-dependent torch.Generator snake_case : List[Any] = self.get_dummy_components() snake_case : Dict = ConsistencyModelPipeline(**A ) snake_case : int = pipe.to(A ) pipe.set_progress_bar_config(disable=A ) snake_case : Any = self.get_dummy_inputs(A ) snake_case : Tuple = pipe(**A ).images assert image.shape == (1, 3_2, 3_2, 3) snake_case : Dict = image[0, -3:, -3:, -1] snake_case : Tuple = np.array([0.35_72, 0.62_73, 0.40_31, 0.39_61, 0.43_21, 0.57_30, 0.52_66, 0.47_80, 0.50_04] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCAmelCase ( self ) -> List[Any]: snake_case : Optional[int] = """cpu""" # ensure determinism for the device-dependent torch.Generator snake_case : Tuple = self.get_dummy_components(class_cond=A ) snake_case : Dict = ConsistencyModelPipeline(**A ) snake_case : List[Any] = pipe.to(A ) pipe.set_progress_bar_config(disable=A ) snake_case : str = self.get_dummy_inputs(A ) snake_case : Optional[int] = 0 snake_case : List[Any] = pipe(**A ).images assert image.shape == (1, 3_2, 3_2, 3) snake_case : Optional[int] = image[0, -3:, -3:, -1] snake_case : Dict = np.array([0.35_72, 0.62_73, 0.40_31, 0.39_61, 0.43_21, 0.57_30, 0.52_66, 0.47_80, 0.50_04] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCAmelCase ( self ) -> Dict: snake_case : str = """cpu""" # ensure determinism for the device-dependent torch.Generator snake_case : Optional[int] = self.get_dummy_components() snake_case : List[str] = ConsistencyModelPipeline(**A ) snake_case : Dict = pipe.to(A ) pipe.set_progress_bar_config(disable=A ) snake_case : Dict = self.get_dummy_inputs(A ) snake_case : Tuple = 1 snake_case : Optional[Any] = None snake_case : List[Any] = pipe(**A ).images assert image.shape == (1, 3_2, 3_2, 3) snake_case : Optional[Any] = image[0, -3:, -3:, -1] snake_case : List[str] = np.array([0.50_04, 0.50_04, 0.49_94, 0.50_08, 0.49_76, 0.50_18, 0.49_90, 0.49_82, 0.49_87] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCAmelCase ( self ) -> Optional[Any]: snake_case : Optional[int] = """cpu""" # ensure determinism for the device-dependent torch.Generator snake_case : Dict = self.get_dummy_components(class_cond=A ) snake_case : List[Any] = ConsistencyModelPipeline(**A ) snake_case : List[Any] = pipe.to(A ) pipe.set_progress_bar_config(disable=A ) snake_case : Optional[Any] = self.get_dummy_inputs(A ) snake_case : Optional[Any] = 1 snake_case : Any = None snake_case : Optional[int] = 0 snake_case : List[Any] = pipe(**A ).images assert image.shape == (1, 3_2, 3_2, 3) snake_case : Optional[int] = image[0, -3:, -3:, -1] snake_case : List[str] = np.array([0.50_04, 0.50_04, 0.49_94, 0.50_08, 0.49_76, 0.50_18, 0.49_90, 0.49_82, 0.49_87] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 @slow @require_torch_gpu class __lowercase (unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> Dict: super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase ( self , A=0 , A=False , A="cpu" , A=torch.floataa , A=(1, 3, 6_4, 6_4) ) -> int: snake_case : Union[str, Any] = torch.manual_seed(A ) snake_case : Tuple = { """num_inference_steps""": None, """timesteps""": [2_2, 0], """class_labels""": 0, """generator""": generator, """output_type""": """np""", } if get_fixed_latents: snake_case : Optional[int] = self.get_fixed_latents(seed=A , device=A , dtype=A , shape=A ) snake_case : int = latents return inputs def UpperCAmelCase ( self , A=0 , A="cpu" , A=torch.floataa , A=(1, 3, 6_4, 6_4) ) -> Any: if type(A ) == str: snake_case : List[str] = torch.device(A ) snake_case : Any = torch.Generator(device=A ).manual_seed(A ) snake_case : Dict = randn_tensor(A , generator=A , device=A , dtype=A ) return latents def UpperCAmelCase ( self ) -> Dict: snake_case : List[Any] = UNetaDModel.from_pretrained("""diffusers/consistency_models""" , subfolder="""diffusers_cd_imagenet64_l2""" ) snake_case : Optional[int] = CMStochasticIterativeScheduler( num_train_timesteps=4_0 , sigma_min=0.0_02 , sigma_max=80.0 , ) snake_case : Any = ConsistencyModelPipeline(unet=A , scheduler=A ) pipe.to(torch_device=A ) pipe.set_progress_bar_config(disable=A ) snake_case : Tuple = self.get_inputs() snake_case : List[str] = pipe(**A ).images assert image.shape == (1, 6_4, 6_4, 3) snake_case : Optional[Any] = image[0, -3:, -3:, -1] snake_case : Union[str, Any] = np.array([0.08_88, 0.08_81, 0.06_66, 0.04_79, 0.02_92, 0.01_95, 0.02_01, 0.01_63, 0.02_54] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 def UpperCAmelCase ( self ) -> Tuple: snake_case : Tuple = UNetaDModel.from_pretrained("""diffusers/consistency_models""" , subfolder="""diffusers_cd_imagenet64_l2""" ) snake_case : int = CMStochasticIterativeScheduler( num_train_timesteps=4_0 , sigma_min=0.0_02 , sigma_max=80.0 , ) snake_case : List[str] = ConsistencyModelPipeline(unet=A , scheduler=A ) pipe.to(torch_device=A ) pipe.set_progress_bar_config(disable=A ) snake_case : Union[str, Any] = self.get_inputs() snake_case : Tuple = 1 snake_case : List[str] = None snake_case : Any = pipe(**A ).images assert image.shape == (1, 6_4, 6_4, 3) snake_case : List[Any] = image[0, -3:, -3:, -1] snake_case : Union[str, Any] = np.array([0.03_40, 0.01_52, 0.00_63, 0.02_67, 0.02_21, 0.01_07, 0.04_16, 0.01_86, 0.02_17] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 @require_torch_a def UpperCAmelCase ( self ) -> Optional[int]: snake_case : List[Any] = UNetaDModel.from_pretrained("""diffusers/consistency_models""" , subfolder="""diffusers_cd_imagenet64_l2""" ) snake_case : Optional[Any] = CMStochasticIterativeScheduler( num_train_timesteps=4_0 , sigma_min=0.0_02 , sigma_max=80.0 , ) snake_case : List[Any] = ConsistencyModelPipeline(unet=A , scheduler=A ) pipe.to(torch_device=A , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=A ) snake_case : List[str] = self.get_inputs(get_fixed_latents=A , device=A ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=A , enable_math=A , enable_mem_efficient=A ): snake_case : Tuple = pipe(**A ).images assert image.shape == (1, 6_4, 6_4, 3) snake_case : Union[str, Any] = image[0, -3:, -3:, -1] snake_case : int = np.array([0.18_75, 0.14_28, 0.12_89, 0.21_51, 0.20_92, 0.14_77, 0.18_77, 0.16_41, 0.13_53] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 @require_torch_a def UpperCAmelCase ( self ) -> Optional[Any]: snake_case : Optional[Any] = UNetaDModel.from_pretrained("""diffusers/consistency_models""" , subfolder="""diffusers_cd_imagenet64_l2""" ) snake_case : Any = CMStochasticIterativeScheduler( num_train_timesteps=4_0 , sigma_min=0.0_02 , sigma_max=80.0 , ) snake_case : Dict = ConsistencyModelPipeline(unet=A , scheduler=A ) pipe.to(torch_device=A , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=A ) snake_case : Optional[int] = self.get_inputs(get_fixed_latents=A , device=A ) snake_case : Union[str, Any] = 1 snake_case : Dict = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=A , enable_math=A , enable_mem_efficient=A ): snake_case : List[Any] = pipe(**A ).images assert image.shape == (1, 6_4, 6_4, 3) snake_case : List[Any] = image[0, -3:, -3:, -1] snake_case : Dict = np.array([0.16_63, 0.19_48, 0.22_75, 0.16_80, 0.12_04, 0.12_45, 0.18_58, 0.13_38, 0.20_95] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3
176
1
def snake_case( __magic_name__ ) -> int: '''simple docstring''' lowercase : List[Any] = abs(__magic_name__ ) lowercase : Optional[Any] = 0 while n > 0: res += n % 10 n //= 10 return res def snake_case( __magic_name__ ) -> int: '''simple docstring''' lowercase : Optional[int] = abs(__magic_name__ ) return n if n < 10 else n % 10 + sum_of_digits(n // 10 ) def snake_case( __magic_name__ ) -> int: '''simple docstring''' return sum(int(__magic_name__ ) for c in str(abs(__magic_name__ ) ) ) def snake_case( ) -> None: '''simple docstring''' from collections.abc import Callable from timeit import timeit def benchmark_a_function(__magic_name__ , __magic_name__ ) -> None: lowercase : str = F"""{func.__name__}({value})""" lowercase : Any = timeit(F"""__main__.{call}""" , setup='''import __main__''' ) print(F"""{call:56} = {func(__magic_name__ )} -- {timing:.4f} seconds""" ) for value in (26_21_44, 11_25_89_99_06_84_26_24, 1_26_76_50_60_02_28_22_94_01_49_67_03_20_53_76): for func in (sum_of_digits, sum_of_digits_recursion, sum_of_digits_compact): benchmark_a_function(__magic_name__ , __magic_name__ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
308
import enum import warnings from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING lowerCAmelCase_ = logging.get_logger(__name__) class _A ( enum.Enum ): _UpperCamelCase : Union[str, Any] = 0 _UpperCamelCase : Any = 1 @add_end_docstrings(_lowerCamelCase ) class _A ( _lowerCamelCase ): _UpperCamelCase : List[Any] = '''generated''' def __init__( self : str , *_A : int , **_A : str ) -> Union[str, Any]: """simple docstring""" super().__init__(*_A , **_A ) self.check_model_type( TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if self.framework == '''tf''' else MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING ) def __a ( self : int , _A : Union[str, Any]=None , _A : Optional[Any]=None , _A : Dict=None , _A : Dict=None , _A : Union[str, Any]=None , _A : int=None , **_A : Optional[int] , ) -> List[Any]: """simple docstring""" lowercase : str = {} if truncation is not None: lowercase : Tuple = truncation lowercase : Tuple = generate_kwargs lowercase : Optional[Any] = {} if return_tensors is not None and return_type is None: lowercase : int = ReturnType.TENSORS if return_tensors else ReturnType.TEXT if return_type is not None: lowercase : Dict = return_type if clean_up_tokenization_spaces is not None: lowercase : Dict = clean_up_tokenization_spaces if stop_sequence is not None: lowercase : Dict = self.tokenizer.encode(_A , add_special_tokens=_A ) if len(_A ) > 1: warnings.warn( '''Stopping on a multiple token sequence is not yet supported on transformers. The first token of''' ''' the stop sequence will be used as the stop sequence string in the interim.''' ) lowercase : List[str] = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def __a ( self : str , _A : int , _A : int , _A : int ) -> List[Any]: """simple docstring""" return True def __a ( self : Union[str, Any] , *_A : Union[str, Any] , _A : List[Any] ) -> Dict: """simple docstring""" lowercase : Tuple = self.model.config.prefix if self.model.config.prefix is not None else '''''' if isinstance(args[0] , _A ): if self.tokenizer.pad_token_id is None: raise ValueError('''Please make sure that the tokenizer has a pad_token_id when using a batch input''' ) lowercase : List[Any] = ([prefix + arg for arg in args[0]],) lowercase : Dict = True elif isinstance(args[0] , _A ): lowercase : Optional[int] = (prefix + args[0],) lowercase : Union[str, Any] = False else: raise ValueError( f""" `args[0]`: {args[0]} have the wrong format. The should be either of type `str` or type `list`""" ) lowercase : Any = self.tokenizer(*_A , padding=_A , truncation=_A , return_tensors=self.framework ) # This is produced by tokenizers but is an invalid generate kwargs if "token_type_ids" in inputs: del inputs["token_type_ids"] return inputs def __call__( self : Union[str, Any] , *_A : Optional[int] , **_A : Tuple ) -> Union[str, Any]: """simple docstring""" lowercase : Any = super().__call__(*_A , **_A ) if ( isinstance(args[0] , _A ) and all(isinstance(_A , _A ) for el in args[0] ) and all(len(_A ) == 1 for res in result ) ): return [res[0] for res in result] return result def __a ( self : Optional[Any] , _A : Optional[Any] , _A : Union[str, Any]=TruncationStrategy.DO_NOT_TRUNCATE , **_A : List[str] ) -> List[Any]: """simple docstring""" lowercase : Optional[int] = self._parse_and_tokenize(_A , truncation=_A , **_A ) return inputs def __a ( self : int , _A : Optional[Any] , **_A : Any ) -> Any: """simple docstring""" if self.framework == "pt": lowercase , lowercase : List[Any] = model_inputs['''input_ids'''].shape elif self.framework == "tf": lowercase , lowercase : Optional[Any] = tf.shape(model_inputs['''input_ids'''] ).numpy() lowercase : int = generate_kwargs.get('''min_length''' , self.model.config.min_length ) lowercase : Optional[int] = generate_kwargs.get('''max_length''' , self.model.config.max_length ) self.check_inputs(_A , generate_kwargs['''min_length'''] , generate_kwargs['''max_length'''] ) lowercase : int = self.model.generate(**_A , **_A ) lowercase : int = output_ids.shape[0] if self.framework == "pt": lowercase : Optional[Any] = output_ids.reshape(_A , out_b // in_b , *output_ids.shape[1:] ) elif self.framework == "tf": lowercase : Tuple = tf.reshape(_A , (in_b, out_b // in_b, *output_ids.shape[1:]) ) return {"output_ids": output_ids} def __a ( self : Union[str, Any] , _A : str , _A : Optional[int]=ReturnType.TEXT , _A : Optional[int]=False ) -> Tuple: """simple docstring""" lowercase : Any = [] for output_ids in model_outputs["output_ids"][0]: if return_type == ReturnType.TENSORS: lowercase : Union[str, Any] = {f"""{self.return_name}_token_ids""": output_ids} elif return_type == ReturnType.TEXT: lowercase : Dict = { f"""{self.return_name}_text""": self.tokenizer.decode( _A , skip_special_tokens=_A , clean_up_tokenization_spaces=_A , ) } records.append(_A ) return records @add_end_docstrings(_lowerCamelCase ) class _A ( _lowerCamelCase ): _UpperCamelCase : List[str] = '''summary''' def __call__( self : List[Any] , *_A : List[str] , **_A : Union[str, Any] ) -> Optional[int]: """simple docstring""" return super().__call__(*_A , **_A ) def __a ( self : Any , _A : int , _A : int , _A : int ) -> bool: """simple docstring""" if max_length < min_length: logger.warning(f"""Your min_length={min_length} must be inferior than your max_length={max_length}.""" ) if input_length < max_length: logger.warning( f"""Your max_length is set to {max_length}, but your input_length is only {input_length}. Since this is """ '''a summarization task, where outputs shorter than the input are typically wanted, you might ''' f"""consider decreasing max_length manually, e.g. summarizer('...', max_length={input_length//2})""" ) @add_end_docstrings(_lowerCamelCase ) class _A ( _lowerCamelCase ): _UpperCamelCase : List[str] = '''translation''' def __a ( self : Union[str, Any] , _A : int , _A : int , _A : int ) -> List[Any]: """simple docstring""" if input_length > 0.9 * max_length: logger.warning( f"""Your input_length: {input_length} is bigger than 0.9 * max_length: {max_length}. You might consider """ '''increasing your max_length manually, e.g. translator(\'...\', max_length=400)''' ) return True def __a ( self : Optional[Any] , *_A : Optional[Any] , _A : Optional[int]=TruncationStrategy.DO_NOT_TRUNCATE , _A : List[Any]=None , _A : Any=None ) -> Dict: """simple docstring""" if getattr(self.tokenizer , '''_build_translation_inputs''' , _A ): return self.tokenizer._build_translation_inputs( *_A , return_tensors=self.framework , truncation=_A , src_lang=_A , tgt_lang=_A ) else: return super()._parse_and_tokenize(*_A , truncation=_A ) def __a ( self : Any , _A : Tuple=None , _A : Any=None , **_A : Any ) -> Optional[int]: """simple docstring""" lowercase , lowercase , lowercase : Dict = super()._sanitize_parameters(**_A ) if src_lang is not None: lowercase : Optional[Any] = src_lang if tgt_lang is not None: lowercase : Dict = tgt_lang if src_lang is None and tgt_lang is None: # Backward compatibility, direct arguments use is preferred. lowercase : Dict = kwargs.get('''task''' , self.task ) lowercase : List[str] = task.split('''_''' ) if task and len(_A ) == 4: # translation, XX, to YY lowercase : Any = items[1] lowercase : List[str] = items[3] return preprocess_params, forward_params, postprocess_params def __call__( self : Tuple , *_A : Union[str, Any] , **_A : List[Any] ) -> List[Any]: """simple docstring""" return super().__call__(*_A , **_A )
308
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from accelerate import PartialState from accelerate.utils.operations import broadcast, gather, gather_object, pad_across_processes, reduce def a_ ( lowerCAmelCase_ : List[str] ): return (torch.arange(state.num_processes ) + 1.0 + (state.num_processes * state.process_index)).to(state.device ) def a_ ( lowerCAmelCase_ : List[str] ): __lowerCAmelCase = create_tensor(lowercase__ ) __lowerCAmelCase = gather(lowercase__ ) assert gathered_tensor.tolist() == list(range(1, state.num_processes**2 + 1 ) ) def a_ ( lowerCAmelCase_ : Optional[int] ): __lowerCAmelCase = [state.process_index] __lowerCAmelCase = gather_object(lowercase__ ) assert len(lowercase__ ) == state.num_processes, F"""{gathered_obj}, {len(lowercase__ )} != {state.num_processes}""" assert gathered_obj == list(range(state.num_processes ) ), F"""{gathered_obj} != {list(range(state.num_processes ) )}""" def a_ ( lowerCAmelCase_ : int ): __lowerCAmelCase = create_tensor(lowercase__ ) __lowerCAmelCase = broadcast(lowercase__ ) assert broadcasted_tensor.shape == torch.Size([state.num_processes] ) assert broadcasted_tensor.tolist() == list(range(1, state.num_processes + 1 ) ) def a_ ( lowerCAmelCase_ : Tuple ): if state.is_main_process: __lowerCAmelCase = torch.arange(state.num_processes + 1 ).to(state.device ) else: __lowerCAmelCase = torch.arange(state.num_processes ).to(state.device ) __lowerCAmelCase = pad_across_processes(lowercase__ ) assert padded_tensor.shape == torch.Size([state.num_processes + 1] ) if not state.is_main_process: assert padded_tensor.tolist() == list(range(0, state.num_processes ) ) + [0] def a_ ( lowerCAmelCase_ : Tuple ): if state.num_processes != 2: return __lowerCAmelCase = create_tensor(lowercase__ ) __lowerCAmelCase = reduce(lowercase__, 'sum' ) __lowerCAmelCase = torch.tensor([4.0, 6] ).to(state.device ) assert torch.allclose(lowercase__, lowercase__ ), F"""{reduced_tensor} != {truth_tensor}""" def a_ ( lowerCAmelCase_ : int ): if state.num_processes != 2: return __lowerCAmelCase = create_tensor(lowercase__ ) __lowerCAmelCase = reduce(lowercase__, 'mean' ) __lowerCAmelCase = torch.tensor([2.0, 3] ).to(state.device ) assert torch.allclose(lowercase__, lowercase__ ), F"""{reduced_tensor} != {truth_tensor}""" def a_ ( lowerCAmelCase_ : Dict ): main() def a_ ( ): __lowerCAmelCase = PartialState() state.print(F"""State: {state}""" ) state.print('testing gather' ) test_gather(lowercase__ ) state.print('testing gather_object' ) test_gather_object(lowercase__ ) state.print('testing broadcast' ) test_broadcast(lowercase__ ) state.print('testing pad_across_processes' ) test_pad_across_processes(lowercase__ ) state.print('testing reduce_sum' ) test_reduce_sum(lowercase__ ) state.print('testing reduce_mean' ) test_reduce_mean(lowercase__ ) if __name__ == "__main__": main()
356
import unittest import numpy as np import torch from .utils_summarization import build_mask, compute_token_type_ids, process_story, truncate_or_pad class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase ( self : Any ) -> Optional[int]: __lowerCAmelCase = 1_0 def lowercase ( self : int ) -> Union[str, Any]: __lowerCAmelCase = [1, 2, 3, 4] __lowerCAmelCase = [1, 2, 3, 4, 0, 0, 0, 0, 0, 0] self.assertEqual(truncate_or_pad(lowerCAmelCase_ , self.block_size , 0 ) , lowerCAmelCase_ ) def lowercase ( self : Optional[Any] ) -> List[str]: __lowerCAmelCase = [1, 2, 3, 4, 5, 6, 7, 8, 9, 1_0] __lowerCAmelCase = [1, 2, 3, 4, 5, 6, 7, 8, 9, 1_0] self.assertEqual(truncate_or_pad(lowerCAmelCase_ , self.block_size , 0 ) , lowerCAmelCase_ ) def lowercase ( self : Any ) -> Optional[Any]: __lowerCAmelCase = [1, 2, 3, 4, 5, 6, 7, 8, 9, 1_0, 1_1, 1_2, 1_3] __lowerCAmelCase = [1, 2, 3, 4, 5, 6, 7, 8, 9, 1_0] self.assertEqual(truncate_or_pad(lowerCAmelCase_ , self.block_size , 0 ) , lowerCAmelCase_ ) def lowercase ( self : List[str] ) -> Any: __lowerCAmelCase = 'It was the year of Our Lord one thousand seven hundred and\n seventy-five.\n\nSpiritual revelations were conceded to England at that\n favoured period, as at this.' __lowerCAmelCase , __lowerCAmelCase = process_story(lowerCAmelCase_ ) self.assertEqual(lowerCAmelCase_ , [] ) def lowercase ( self : Any ) -> str: __lowerCAmelCase = '' __lowerCAmelCase , __lowerCAmelCase = process_story(lowerCAmelCase_ ) self.assertEqual(lowerCAmelCase_ , [] ) self.assertEqual(lowerCAmelCase_ , [] ) def lowercase ( self : int ) -> int: __lowerCAmelCase = ( 'It was the year of Our Lord one thousand seven hundred and ' 'seventy-five\n\nSpiritual revelations were conceded to England ' 'at that favoured period, as at this.\n@highlight\n\nIt was the best of times' ) __lowerCAmelCase , __lowerCAmelCase = process_story(lowerCAmelCase_ ) __lowerCAmelCase = [ 'It was the year of Our Lord one thousand seven hundred and seventy-five.', 'Spiritual revelations were conceded to England at that favoured period, as at this.', ] self.assertEqual(lowerCAmelCase_ , lowerCAmelCase_ ) __lowerCAmelCase = ['It was the best of times.'] self.assertEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def lowercase ( self : Dict ) -> Any: __lowerCAmelCase = torch.tensor([1, 2, 3, 4] ) __lowerCAmelCase = torch.tensor([1, 1, 1, 1] ) np.testing.assert_array_equal(build_mask(lowerCAmelCase_ , 0 ).numpy() , expected.numpy() ) def lowercase ( self : List[Any] ) -> Optional[int]: __lowerCAmelCase = torch.tensor([1, 2, 3, 4, 2_3, 2_3, 2_3] ) __lowerCAmelCase = torch.tensor([1, 1, 1, 1, 0, 0, 0] ) np.testing.assert_array_equal(build_mask(lowerCAmelCase_ , 2_3 ).numpy() , expected.numpy() ) def lowercase ( self : str ) -> List[Any]: __lowerCAmelCase = torch.tensor([8, 2, 3, 4, 1, 1, 1] ) __lowerCAmelCase = torch.tensor([1, 1, 1, 1, 0, 0, 0] ) np.testing.assert_array_equal(build_mask(lowerCAmelCase_ , 1 ).numpy() , expected.numpy() ) def lowercase ( self : Optional[Any] ) -> Optional[int]: __lowerCAmelCase = 1_0_1 __lowerCAmelCase = torch.tensor([[1, 2, 3, 4, 5, 6], [1, 2, 3, 1_0_1, 5, 6], [1, 1_0_1, 3, 4, 1_0_1, 6]] ) __lowerCAmelCase = torch.tensor([[1, 1, 1, 1, 1, 1], [1, 1, 1, 0, 0, 0], [1, 0, 0, 0, 1, 1]] ) __lowerCAmelCase = compute_token_type_ids(lowerCAmelCase_ , lowerCAmelCase_ ) np.testing.assert_array_equal(lowerCAmelCase_ , lowerCAmelCase_ )
207
0
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyInpaintPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.text_encoder import MCLIPConfig, MultilingualCLIP from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class A__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = KandinskyInpaintPipeline SCREAMING_SNAKE_CASE = ['prompt', 'image_embeds', 'negative_image_embeds', 'image', 'mask_image'] SCREAMING_SNAKE_CASE = [ 'prompt', 'negative_prompt', 'image_embeds', 'negative_image_embeds', 'image', 'mask_image', ] SCREAMING_SNAKE_CASE = [ 'generator', 'height', 'width', 'latents', 'guidance_scale', 'negative_prompt', 'num_inference_steps', 'return_dict', 'guidance_scale', 'num_images_per_prompt', 'output_type', 'return_dict', ] SCREAMING_SNAKE_CASE = False @property def _SCREAMING_SNAKE_CASE ( self: List[Any]) -> Optional[Any]: """simple docstring""" return 32 @property def _SCREAMING_SNAKE_CASE ( self: Dict) -> Optional[int]: """simple docstring""" return 32 @property def _SCREAMING_SNAKE_CASE ( self: Any) -> Dict: """simple docstring""" return self.time_input_dim @property def _SCREAMING_SNAKE_CASE ( self: int) -> Union[str, Any]: """simple docstring""" return self.time_input_dim * 4 @property def _SCREAMING_SNAKE_CASE ( self: Tuple) -> str: """simple docstring""" return 100 @property def _SCREAMING_SNAKE_CASE ( self: Tuple) -> Dict: """simple docstring""" __lowerCAmelCase : Optional[Any] = XLMRobertaTokenizerFast.from_pretrained("YiYiXu/tiny-random-mclip-base") return tokenizer @property def _SCREAMING_SNAKE_CASE ( self: Dict) -> Optional[Any]: """simple docstring""" torch.manual_seed(0) __lowerCAmelCase : str = MCLIPConfig( numDims=self.cross_attention_dim , transformerDimensions=self.text_embedder_hidden_size , hidden_size=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=1005 , ) __lowerCAmelCase : int = MultilingualCLIP(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = text_encoder.eval() return text_encoder @property def _SCREAMING_SNAKE_CASE ( self: Optional[int]) -> List[str]: """simple docstring""" torch.manual_seed(0) __lowerCAmelCase : str = { "in_channels": 9, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "text_image", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "text_image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } __lowerCAmelCase : Any = UNetaDConditionModel(**_SCREAMING_SNAKE_CASE) return model @property def _SCREAMING_SNAKE_CASE ( self: str) -> Optional[Any]: """simple docstring""" return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def _SCREAMING_SNAKE_CASE ( self: Dict) -> Optional[Any]: """simple docstring""" torch.manual_seed(0) __lowerCAmelCase : List[str] = VQModel(**self.dummy_movq_kwargs) return model def _SCREAMING_SNAKE_CASE ( self: int) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Union[str, Any] = self.dummy_text_encoder __lowerCAmelCase : Optional[Any] = self.dummy_tokenizer __lowerCAmelCase : Union[str, Any] = self.dummy_unet __lowerCAmelCase : Any = self.dummy_movq __lowerCAmelCase : int = DDIMScheduler( num_train_timesteps=1000 , beta_schedule="linear" , beta_start=0.0_0085 , beta_end=0.012 , clip_sample=_SCREAMING_SNAKE_CASE , set_alpha_to_one=_SCREAMING_SNAKE_CASE , steps_offset=1 , prediction_type="epsilon" , thresholding=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : List[str] = { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "movq": movq, } return components def _SCREAMING_SNAKE_CASE ( self: Tuple , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: str=0) -> Optional[int]: """simple docstring""" __lowerCAmelCase : Tuple = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(_SCREAMING_SNAKE_CASE)).to(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Union[str, Any] = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(seed + 1)).to(_SCREAMING_SNAKE_CASE) # create init_image __lowerCAmelCase : Optional[Any] = floats_tensor((1, 3, 64, 64) , rng=random.Random(_SCREAMING_SNAKE_CASE)).to(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = image.cpu().permute(0 , 2 , 3 , 1)[0] __lowerCAmelCase : List[str] = Image.fromarray(np.uinta(_SCREAMING_SNAKE_CASE)).convert("RGB").resize((256, 256)) # create mask __lowerCAmelCase : List[str] = np.ones((64, 64) , dtype=np.floataa) __lowerCAmelCase : Optional[Any] = 0 if str(_SCREAMING_SNAKE_CASE).startswith("mps"): __lowerCAmelCase : Tuple = torch.manual_seed(_SCREAMING_SNAKE_CASE) else: __lowerCAmelCase : str = torch.Generator(device=_SCREAMING_SNAKE_CASE).manual_seed(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : str = { "prompt": "horse", "image": init_image, "mask_image": mask, "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "generator": generator, "height": 64, "width": 64, "num_inference_steps": 2, "guidance_scale": 4.0, "output_type": "np", } return inputs def _SCREAMING_SNAKE_CASE ( self: Any) -> List[Any]: """simple docstring""" __lowerCAmelCase : str = "cpu" __lowerCAmelCase : int = self.get_dummy_components() __lowerCAmelCase : Optional[int] = self.pipeline_class(**_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = pipe.to(_SCREAMING_SNAKE_CASE) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = pipe(**self.get_dummy_inputs(_SCREAMING_SNAKE_CASE)) __lowerCAmelCase : List[Any] = output.images __lowerCAmelCase : List[Any] = pipe( **self.get_dummy_inputs(_SCREAMING_SNAKE_CASE) , return_dict=_SCREAMING_SNAKE_CASE , )[0] __lowerCAmelCase : str = image[0, -3:, -3:, -1] __lowerCAmelCase : int = image_from_tuple[0, -3:, -3:, -1] print(F"""image.shape {image.shape}""") assert image.shape == (1, 64, 64, 3) __lowerCAmelCase : Optional[Any] = np.array( [0.832_6919, 0.7379_0467, 0.2091_8581, 0.930_9612, 0.551_1791, 0.4371_3328, 0.551_3321, 0.4992_2934, 0.5949_7786]) assert ( np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1e-2 ), F""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" def _SCREAMING_SNAKE_CASE ( self: Any) -> Union[str, Any]: """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3e-3) @slow @require_torch_gpu class A__ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: str) -> List[str]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _SCREAMING_SNAKE_CASE ( self: Union[str, Any]) -> List[str]: """simple docstring""" __lowerCAmelCase : List[str] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinsky/kandinsky_inpaint_cat_with_hat_fp16.npy") __lowerCAmelCase : Any = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinsky/cat.png") __lowerCAmelCase : List[Any] = np.ones((768, 768) , dtype=np.floataa) __lowerCAmelCase : List[str] = 0 __lowerCAmelCase : Dict = "a hat" __lowerCAmelCase : Optional[Any] = KandinskyPriorPipeline.from_pretrained( "kandinsky-community/kandinsky-2-1-prior" , torch_dtype=torch.floataa) pipe_prior.to(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = KandinskyInpaintPipeline.from_pretrained( "kandinsky-community/kandinsky-2-1-inpaint" , torch_dtype=torch.floataa) __lowerCAmelCase : Optional[int] = pipeline.to(_SCREAMING_SNAKE_CASE) pipeline.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = torch.Generator(device="cpu").manual_seed(0) __lowerCAmelCase , __lowerCAmelCase : Union[str, Any] = pipe_prior( _SCREAMING_SNAKE_CASE , generator=_SCREAMING_SNAKE_CASE , num_inference_steps=5 , negative_prompt="" , ).to_tuple() __lowerCAmelCase : Optional[int] = pipeline( _SCREAMING_SNAKE_CASE , image=_SCREAMING_SNAKE_CASE , mask_image=_SCREAMING_SNAKE_CASE , image_embeds=_SCREAMING_SNAKE_CASE , negative_image_embeds=_SCREAMING_SNAKE_CASE , generator=_SCREAMING_SNAKE_CASE , num_inference_steps=100 , height=768 , width=768 , output_type="np" , ) __lowerCAmelCase : Union[str, Any] = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE)
269
"""simple docstring""" from typing import TYPE_CHECKING from ..utils import _LazyModule __snake_case : Optional[int] = { 'config': [ 'EXTERNAL_DATA_FORMAT_SIZE_LIMIT', 'OnnxConfig', 'OnnxConfigWithPast', 'OnnxSeq2SeqConfigWithPast', 'PatchingSpec', ], 'convert': ['export', 'validate_model_outputs'], 'features': ['FeaturesManager'], 'utils': ['ParameterFormat', 'compute_serialized_parameters_size'], } if TYPE_CHECKING: from .config import ( EXTERNAL_DATA_FORMAT_SIZE_LIMIT, OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast, PatchingSpec, ) from .convert import export, validate_model_outputs from .features import FeaturesManager from .utils import ParameterFormat, compute_serialized_parameters_size else: import sys __snake_case : Any = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
269
1
'''simple docstring''' from ..utils import DummyObject, requires_backends class _A ( metaclass=__lowercase ): lowercase__: Tuple = ['''torch''', '''transformers''', '''onnx'''] def __init__( self : str , *__magic_name__ : Union[str, Any] , **__magic_name__ : Optional[int] ) -> Optional[Any]: """simple docstring""" requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def lowercase__ ( cls : Optional[Any] , *__magic_name__ : Union[str, Any] , **__magic_name__ : int ) -> str: """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def lowercase__ ( cls : List[Any] , *__magic_name__ : Dict , **__magic_name__ : Dict ) -> List[str]: """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) class _A ( metaclass=__lowercase ): lowercase__: Tuple = ['''torch''', '''transformers''', '''onnx'''] def __init__( self : Tuple , *__magic_name__ : Any , **__magic_name__ : Optional[Any] ) -> Optional[Any]: """simple docstring""" requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def lowercase__ ( cls : Tuple , *__magic_name__ : str , **__magic_name__ : Tuple ) -> int: """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def lowercase__ ( cls : Optional[int] , *__magic_name__ : Optional[int] , **__magic_name__ : int ) -> List[str]: """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) class _A ( metaclass=__lowercase ): lowercase__: Union[str, Any] = ['''torch''', '''transformers''', '''onnx'''] def __init__( self : int , *__magic_name__ : int , **__magic_name__ : Union[str, Any] ) -> List[str]: """simple docstring""" requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def lowercase__ ( cls : int , *__magic_name__ : str , **__magic_name__ : List[str] ) -> Union[str, Any]: """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def lowercase__ ( cls : int , *__magic_name__ : Dict , **__magic_name__ : Any ) -> str: """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) class _A ( metaclass=__lowercase ): lowercase__: Tuple = ['''torch''', '''transformers''', '''onnx'''] def __init__( self : List[Any] , *__magic_name__ : List[str] , **__magic_name__ : Union[str, Any] ) -> int: """simple docstring""" requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def lowercase__ ( cls : Dict , *__magic_name__ : List[str] , **__magic_name__ : Dict ) -> List[str]: """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def lowercase__ ( cls : Optional[Any] , *__magic_name__ : Optional[int] , **__magic_name__ : List[str] ) -> Optional[int]: """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) class _A ( metaclass=__lowercase ): lowercase__: Optional[Any] = ['''torch''', '''transformers''', '''onnx'''] def __init__( self : Optional[int] , *__magic_name__ : str , **__magic_name__ : Dict ) -> Optional[int]: """simple docstring""" requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def lowercase__ ( cls : str , *__magic_name__ : Any , **__magic_name__ : List[Any] ) -> Dict: """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def lowercase__ ( cls : Any , *__magic_name__ : Tuple , **__magic_name__ : str ) -> str: """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) class _A ( metaclass=__lowercase ): lowercase__: Dict = ['''torch''', '''transformers''', '''onnx'''] def __init__( self : int , *__magic_name__ : Optional[Any] , **__magic_name__ : Optional[int] ) -> Tuple: """simple docstring""" requires_backends(self , ["""torch""", """transformers""", """onnx"""] ) @classmethod def lowercase__ ( cls : Any , *__magic_name__ : Tuple , **__magic_name__ : Any ) -> Any: """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] ) @classmethod def lowercase__ ( cls : Optional[int] , *__magic_name__ : int , **__magic_name__ : Dict ) -> Optional[int]: """simple docstring""" requires_backends(cls , ["""torch""", """transformers""", """onnx"""] )
13
'''simple docstring''' import copy import inspect import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import TimesformerConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, TimesformerForVideoClassification, TimesformerModel, ) from transformers.models.timesformer.modeling_timesformer import TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from transformers import VideoMAEImageProcessor class _A : def __init__( self : Tuple , __magic_name__ : List[str] , __magic_name__ : str=13 , __magic_name__ : int=10 , __magic_name__ : Any=3 , __magic_name__ : List[Any]=2 , __magic_name__ : List[Any]=2 , __magic_name__ : Union[str, Any]=True , __magic_name__ : Union[str, Any]=True , __magic_name__ : Any=32 , __magic_name__ : int=5 , __magic_name__ : Optional[int]=4 , __magic_name__ : List[Any]=37 , __magic_name__ : Dict="gelu" , __magic_name__ : List[Any]=0.1 , __magic_name__ : Optional[int]=0.1 , __magic_name__ : Any=10 , __magic_name__ : List[str]=0.02 , __magic_name__ : Optional[Any]="divided_space_time" , __magic_name__ : int=None , ) -> List[str]: """simple docstring""" __snake_case : List[Any] = parent __snake_case : List[str] = batch_size __snake_case : Union[str, Any] = image_size __snake_case : List[Any] = num_channels __snake_case : List[str] = patch_size __snake_case : List[str] = num_frames __snake_case : Union[str, Any] = is_training __snake_case : List[str] = use_labels __snake_case : str = hidden_size __snake_case : Union[str, Any] = num_hidden_layers __snake_case : Union[str, Any] = num_attention_heads __snake_case : Dict = intermediate_size __snake_case : Tuple = hidden_act __snake_case : Optional[Any] = hidden_dropout_prob __snake_case : Optional[int] = attention_probs_dropout_prob __snake_case : Union[str, Any] = attention_type __snake_case : Optional[Any] = initializer_range __snake_case : Optional[Any] = scope __snake_case : Optional[int] = num_labels # in TimeSformer, the number of spatial tokens equals num_frames * num_patches per frame + 1 CLS token __snake_case : str = (image_size // patch_size) ** 2 __snake_case : Optional[Any] = (num_frames) * self.num_patches_per_frame + 1 def lowercase__ ( self : List[str] ) -> Optional[int]: """simple docstring""" __snake_case : Optional[int] = floats_tensor( [self.batch_size, self.num_frames, self.num_channels, self.image_size, self.image_size] ) __snake_case : int = None if self.use_labels: __snake_case : Optional[Any] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : int = self.get_config() return config, pixel_values, labels def lowercase__ ( self : Any ) -> Union[str, Any]: """simple docstring""" __snake_case : Any = TimesformerConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_frames=self.num_frames , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , attention_type=self.attention_type , ) __snake_case : str = self.num_labels return config def lowercase__ ( self : List[Any] , __magic_name__ : Tuple , __magic_name__ : Tuple , __magic_name__ : Dict ) -> int: """simple docstring""" __snake_case : Optional[int] = TimesformerModel(config=__magic_name__ ) model.to(__magic_name__ ) model.eval() __snake_case : Tuple = model(__magic_name__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase__ ( self : Any , __magic_name__ : Optional[int] , __magic_name__ : str , __magic_name__ : Optional[int] ) -> str: """simple docstring""" __snake_case : Any = TimesformerForVideoClassification(__magic_name__ ) model.to(__magic_name__ ) model.eval() __snake_case : Optional[int] = model(__magic_name__ ) # verify the logits shape __snake_case : Dict = torch.Size((self.batch_size, self.num_labels) ) self.parent.assertEqual(result.logits.shape , __magic_name__ ) def lowercase__ ( self : Optional[Any] ) -> List[Any]: """simple docstring""" __snake_case : Optional[Any] = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case : Tuple = config_and_inputs __snake_case : List[Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _A ( __lowercase , __lowercase , unittest.TestCase ): lowercase__: Dict = (TimesformerModel, TimesformerForVideoClassification) if is_torch_available() else () lowercase__: List[Any] = ( {'''feature-extraction''': TimesformerModel, '''video-classification''': TimesformerForVideoClassification} if is_torch_available() else {} ) lowercase__: List[str] = False lowercase__: List[Any] = False lowercase__: Dict = False lowercase__: int = False def lowercase__ ( self : Any ) -> int: """simple docstring""" __snake_case : List[str] = TimesformerModelTester(self ) __snake_case : List[Any] = ConfigTester( self , config_class=__magic_name__ , has_text_modality=__magic_name__ , hidden_size=37 ) def lowercase__ ( self : Any , __magic_name__ : Tuple , __magic_name__ : List[str] , __magic_name__ : Union[str, Any]=False ) -> int: """simple docstring""" __snake_case : Dict = copy.deepcopy(__magic_name__ ) if return_labels: if model_class in get_values(__magic_name__ ): __snake_case : List[str] = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__magic_name__ ) return inputs_dict def lowercase__ ( self : Tuple ) -> Union[str, Any]: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""TimeSformer does not use inputs_embeds""" ) def lowercase__ ( self : List[str] ) -> Any: """simple docstring""" pass def lowercase__ ( self : str ) -> Optional[int]: """simple docstring""" __snake_case , __snake_case : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : str = model_class(__magic_name__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __snake_case : Tuple = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__magic_name__ , nn.Linear ) ) def lowercase__ ( self : Any ) -> int: """simple docstring""" __snake_case , __snake_case : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Union[str, Any] = model_class(__magic_name__ ) __snake_case : Optional[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : Union[str, Any] = [*signature.parameters.keys()] __snake_case : str = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __magic_name__ ) def lowercase__ ( self : str ) -> Dict: """simple docstring""" __snake_case : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__magic_name__ ) def lowercase__ ( self : int ) -> List[str]: """simple docstring""" __snake_case : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_video_classification(*__magic_name__ ) @slow def lowercase__ ( self : List[Any] ) -> Dict: """simple docstring""" for model_name in TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : int = TimesformerModel.from_pretrained(__magic_name__ ) self.assertIsNotNone(__magic_name__ ) def lowercase__ ( self : Dict ) -> Optional[int]: """simple docstring""" if not self.has_attentions: pass else: __snake_case , __snake_case : Tuple = self.model_tester.prepare_config_and_inputs_for_common() __snake_case : Dict = True for model_class in self.all_model_classes: __snake_case : List[str] = self.model_tester.seq_length __snake_case : Tuple = self.model_tester.num_frames __snake_case : str = True __snake_case : List[str] = False __snake_case : Tuple = True __snake_case : str = model_class(__magic_name__ ) model.to(__magic_name__ ) model.eval() with torch.no_grad(): __snake_case : List[str] = model(**self._prepare_for_class(__magic_name__ , __magic_name__ ) ) __snake_case : Dict = outputs.attentions self.assertEqual(len(__magic_name__ ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] __snake_case : Optional[int] = True __snake_case : Any = model_class(__magic_name__ ) model.to(__magic_name__ ) model.eval() with torch.no_grad(): __snake_case : Union[str, Any] = model(**self._prepare_for_class(__magic_name__ , __magic_name__ ) ) __snake_case : int = outputs.attentions self.assertEqual(len(__magic_name__ ) , self.model_tester.num_hidden_layers ) # attentions has shape (batch_size x num_frames) x num_heads x (num_patches per frame + 1) x (num_patches per frame + 1) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len // num_frames + 1, seq_len // num_frames + 1] , ) __snake_case : int = len(__magic_name__ ) # Check attention is always last and order is fine __snake_case : Optional[int] = True __snake_case : Optional[int] = True __snake_case : Union[str, Any] = model_class(__magic_name__ ) model.to(__magic_name__ ) model.eval() with torch.no_grad(): __snake_case : Dict = model(**self._prepare_for_class(__magic_name__ , __magic_name__ ) ) self.assertEqual(out_len + 1 , len(__magic_name__ ) ) __snake_case : List[Any] = outputs.attentions self.assertEqual(len(__magic_name__ ) , self.model_tester.num_hidden_layers ) # attentions has shape (batch_size x num_frames) x num_heads x (num_patches per frame + 1) x (num_patches per frame + 1) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len // num_frames + 1, seq_len // num_frames + 1] , ) def lowercase__ ( self : Dict ) -> int: """simple docstring""" def check_hidden_states_output(__magic_name__ : List[str] , __magic_name__ : List[str] , __magic_name__ : Optional[Any] ): __snake_case : str = model_class(__magic_name__ ) model.to(__magic_name__ ) model.eval() with torch.no_grad(): __snake_case : Tuple = model(**self._prepare_for_class(__magic_name__ , __magic_name__ ) ) __snake_case : int = outputs.hidden_states __snake_case : Dict = self.model_tester.num_hidden_layers + 1 self.assertEqual(len(__magic_name__ ) , __magic_name__ ) __snake_case : int = self.model_tester.seq_length self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) __snake_case , __snake_case : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Dict = True check_hidden_states_output(__magic_name__ , __magic_name__ , __magic_name__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : str = True check_hidden_states_output(__magic_name__ , __magic_name__ , __magic_name__ ) def _a ( ) -> List[Any]: """simple docstring""" __snake_case : Optional[Any] = hf_hub_download( repo_id="""hf-internal-testing/spaghetti-video""" , filename="""eating_spaghetti.npy""" , repo_type="""dataset""" ) __snake_case : List[Any] = np.load(_lowerCamelCase ) return list(_lowerCamelCase ) @require_torch @require_vision class _A ( unittest.TestCase ): @cached_property def lowercase__ ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" return ( VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5] , image_std=[0.5, 0.5, 0.5] ) if is_vision_available() else None ) @slow def lowercase__ ( self : Optional[int] ) -> List[str]: """simple docstring""" __snake_case : int = TimesformerForVideoClassification.from_pretrained("""facebook/timesformer-base-finetuned-k400""" ).to( __magic_name__ ) __snake_case : Union[str, Any] = self.default_image_processor __snake_case : Dict = prepare_video() __snake_case : Any = image_processor(video[:8] , return_tensors="""pt""" ).to(__magic_name__ ) # forward pass with torch.no_grad(): __snake_case : Any = model(**__magic_name__ ) # verify the logits __snake_case : int = torch.Size((1, 4_00) ) self.assertEqual(outputs.logits.shape , __magic_name__ ) __snake_case : Any = torch.tensor([-0.3016, -0.7713, -0.4205] ).to(__magic_name__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __magic_name__ , atol=1E-4 ) )
13
1
"""simple docstring""" import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskaFormerForUniversalSegmentation, MaskaFormerModel if is_vision_available(): from transformers import MaskaFormerImageProcessor if is_vision_available(): from PIL import Image class UpperCamelCase : def __init__(self : int , _A : str , _A : List[Any]=2 , _A : Tuple=True , _A : Optional[Any]=False , _A : List[Any]=10 , _A : int=3 , _A : Union[str, Any]=32 * 8 , _A : List[str]=32 * 8 , _A : Dict=4 , _A : Any=64 , ) -> List[Any]: __snake_case : Tuple = parent __snake_case : Union[str, Any] = batch_size __snake_case : Any = is_training __snake_case : Tuple = use_auxiliary_loss __snake_case : Optional[Any] = num_queries __snake_case : Optional[Any] = num_channels __snake_case : str = min_size __snake_case : List[Any] = max_size __snake_case : Tuple = num_labels __snake_case : Tuple = hidden_dim __snake_case : Tuple = hidden_dim def _lowercase (self : int) -> List[Any]: __snake_case : Any = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size]).to( _A) __snake_case : Tuple = torch.ones([self.batch_size, self.min_size, self.max_size] , device=_A) __snake_case : Any = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=_A) > 0.5 ).float() __snake_case : Union[str, Any] = (torch.rand((self.batch_size, self.num_labels) , device=_A) > 0.5).long() __snake_case : Dict = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def _lowercase (self : Union[str, Any]) -> Tuple: __snake_case : str = MaskaFormerConfig( hidden_size=self.hidden_dim , ) __snake_case : int = self.num_queries __snake_case : Any = self.num_labels __snake_case : List[Any] = [1, 1, 1, 1] __snake_case : Dict = self.num_channels __snake_case : Any = 64 __snake_case : str = 1_28 __snake_case : Optional[Any] = self.hidden_dim __snake_case : List[Any] = self.hidden_dim __snake_case : Tuple = self.hidden_dim return config def _lowercase (self : Any) -> Dict: __snake_case , __snake_case , __snake_case , __snake_case , __snake_case : List[str] = self.prepare_config_and_inputs() __snake_case : Any = {'pixel_values': pixel_values, 'pixel_mask': pixel_mask} return config, inputs_dict def _lowercase (self : Optional[int] , _A : str , _A : List[Any]) -> List[str]: __snake_case : List[Any] = output.encoder_hidden_states __snake_case : Tuple = output.pixel_decoder_hidden_states __snake_case : str = output.transformer_decoder_hidden_states self.parent.assertTrue(len(_A) , len(config.backbone_config.depths)) self.parent.assertTrue(len(_A) , len(config.backbone_config.depths)) self.parent.assertTrue(len(_A) , config.decoder_layers) def _lowercase (self : Tuple , _A : int , _A : Dict , _A : Optional[int] , _A : Tuple=False) -> Union[str, Any]: with torch.no_grad(): __snake_case : List[Any] = MaskaFormerModel(config=_A) model.to(_A) model.eval() __snake_case : Dict = model(pixel_values=_A , pixel_mask=_A) __snake_case : List[str] = model(_A , output_hidden_states=_A) self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.hidden_dim) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None) self.parent.assertTrue(output.encoder_last_hidden_state is not None) if output_hidden_states: self.check_output_hidden_state(_A , _A) def _lowercase (self : Any , _A : int , _A : Optional[int] , _A : Dict , _A : Union[str, Any] , _A : str) -> List[str]: __snake_case : Any = MaskaFormerForUniversalSegmentation(config=_A) model.to(_A) model.eval() def comm_check_on_output(_A : int): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None) self.parent.assertTrue(result.encoder_last_hidden_state is not None) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1)) with torch.no_grad(): __snake_case : Any = model(pixel_values=_A , pixel_mask=_A) __snake_case : str = model(_A) comm_check_on_output(_A) __snake_case : List[Any] = model( pixel_values=_A , pixel_mask=_A , mask_labels=_A , class_labels=_A) comm_check_on_output(_A) self.parent.assertTrue(result.loss is not None) self.parent.assertEqual(result.loss.shape , torch.Size([1])) @require_torch class UpperCamelCase ( lowercase , lowercase , unittest.TestCase ): UpperCAmelCase : str = (MaskaFormerModel, MaskaFormerForUniversalSegmentation) if is_torch_available() else () UpperCAmelCase : Optional[int] = {"""feature-extraction""": MaskaFormerModel} if is_torch_available() else {} UpperCAmelCase : List[str] = False UpperCAmelCase : List[Any] = False UpperCAmelCase : Union[str, Any] = False UpperCAmelCase : str = False def _lowercase (self : Dict) -> List[Any]: __snake_case : Tuple = MaskaFormerModelTester(self) __snake_case : Optional[Any] = ConfigTester(self , config_class=_A , has_text_modality=_A) def _lowercase (self : Optional[Any]) -> List[str]: self.config_tester.run_common_tests() def _lowercase (self : Any) -> Any: __snake_case , __snake_case : Any = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_A , **_A , output_hidden_states=_A) def _lowercase (self : Tuple) -> List[Any]: __snake_case : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskaformer_instance_segmentation_head_model(*_A) @unittest.skip(reason='Mask2Former does not use inputs_embeds') def _lowercase (self : int) -> Any: pass @unittest.skip(reason='Mask2Former does not have a get_input_embeddings method') def _lowercase (self : Any) -> Union[str, Any]: pass @unittest.skip(reason='Mask2Former is not a generative model') def _lowercase (self : Optional[int]) -> Union[str, Any]: pass @unittest.skip(reason='Mask2Former does not use token embeddings') def _lowercase (self : List[Any]) -> Tuple: pass @require_torch_multi_gpu @unittest.skip( reason='Mask2Former has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`') def _lowercase (self : Optional[Any]) -> Union[str, Any]: pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.') def _lowercase (self : int) -> Union[str, Any]: pass def _lowercase (self : Optional[int]) -> List[Any]: __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Tuple = model_class(_A) __snake_case : str = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : Any = [*signature.parameters.keys()] __snake_case : Optional[Any] = ['pixel_values'] self.assertListEqual(arg_names[:1] , _A) @slow def _lowercase (self : Tuple) -> Any: for model_name in ["facebook/mask2former-swin-small-coco-instance"]: __snake_case : Any = MaskaFormerModel.from_pretrained(_A) self.assertIsNotNone(_A) def _lowercase (self : Dict) -> Tuple: __snake_case : List[Any] = (self.model_tester.min_size,) * 2 __snake_case : List[str] = { 'pixel_values': torch.randn((2, 3, *size) , device=_A), 'mask_labels': torch.randn((2, 10, *size) , device=_A), 'class_labels': torch.zeros(2 , 10 , device=_A).long(), } __snake_case : Dict = self.model_tester.get_config() __snake_case : int = MaskaFormerForUniversalSegmentation(_A).to(_A) __snake_case : int = model(**_A) self.assertTrue(outputs.loss is not None) def _lowercase (self : List[str]) -> List[str]: __snake_case , __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_A , **_A , output_hidden_states=_A) def _lowercase (self : Any) -> List[Any]: __snake_case , __snake_case : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Any = model_class(_A).to(_A) __snake_case : int = model(**_A , output_attentions=_A) self.assertTrue(outputs.attentions is not None) def _lowercase (self : Tuple) -> List[str]: if not self.model_tester.is_training: return __snake_case : Any = self.all_model_classes[1] __snake_case , __snake_case , __snake_case , __snake_case , __snake_case : Union[str, Any] = self.model_tester.prepare_config_and_inputs() __snake_case : Optional[Any] = model_class(_A) model.to(_A) model.train() __snake_case : Optional[int] = model(_A , mask_labels=_A , class_labels=_A).loss loss.backward() def _lowercase (self : Any) -> Union[str, Any]: __snake_case : Union[str, Any] = self.all_model_classes[1] __snake_case , __snake_case , __snake_case , __snake_case , __snake_case : Any = self.model_tester.prepare_config_and_inputs() __snake_case : Optional[Any] = True __snake_case : Dict = True __snake_case : Dict = model_class(_A).to(_A) model.train() __snake_case : List[str] = model(_A , mask_labels=_A , class_labels=_A) __snake_case : Any = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() __snake_case : Optional[int] = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() __snake_case : Optional[int] = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() __snake_case : Optional[Any] = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=_A) self.assertIsNotNone(encoder_hidden_states.grad) self.assertIsNotNone(pixel_decoder_hidden_states.grad) self.assertIsNotNone(transformer_decoder_hidden_states.grad) self.assertIsNotNone(attentions.grad) _a : Any= 1e-4 def __UpperCAmelCase ( ) -> Optional[int]: '''simple docstring''' __snake_case : Tuple = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_vision @slow class UpperCamelCase ( unittest.TestCase ): @cached_property def _lowercase (self : Optional[Any]) -> List[str]: return "facebook/mask2former-swin-small-coco-instance" @cached_property def _lowercase (self : List[Any]) -> Optional[int]: return MaskaFormerImageProcessor.from_pretrained(self.model_checkpoints) if is_vision_available() else None def _lowercase (self : Any) -> Any: __snake_case : Optional[Any] = MaskaFormerModel.from_pretrained(self.model_checkpoints).to(_A) __snake_case : Optional[Any] = self.default_image_processor __snake_case : Dict = prepare_img() __snake_case : Dict = image_processor(_A , return_tensors='pt').to(_A) __snake_case : Union[str, Any] = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0) # check size self.assertEqual(_A , (1, 3, 3_84, 3_84)) with torch.no_grad(): __snake_case : Union[str, Any] = model(**_A) __snake_case : List[Any] = torch.tensor( [[-0.2_790, -1.0_717, -1.1_668], [-0.5_128, -0.3_128, -0.4_987], [-0.5_832, 0.1_971, -0.0_197]]).to(_A) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , _A , atol=_A)) __snake_case : Optional[int] = torch.tensor( [[0.8_973, 1.1_847, 1.1_776], [1.1_934, 1.5_040, 1.5_128], [1.1_153, 1.4_486, 1.4_951]]).to(_A) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , _A , atol=_A)) __snake_case : List[str] = torch.tensor( [[2.1_152, 1.7_000, -0.8_603], [1.5_808, 1.8_004, -0.9_353], [1.6_043, 1.7_495, -0.5_999]]).to(_A) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , _A , atol=_A)) def _lowercase (self : Any) -> str: __snake_case : Dict = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints).to(_A).eval() __snake_case : Optional[Any] = self.default_image_processor __snake_case : str = prepare_img() __snake_case : int = image_processor(_A , return_tensors='pt').to(_A) __snake_case : List[str] = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0) # check size self.assertEqual(_A , (1, 3, 3_84, 3_84)) with torch.no_grad(): __snake_case : Union[str, Any] = model(**_A) # masks_queries_logits __snake_case : List[str] = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4)) __snake_case : str = [ [-8.7_839, -9.0_056, -8.8_121], [-7.4_104, -7.0_313, -6.5_401], [-6.6_105, -6.3_427, -6.4_675], ] __snake_case : Any = torch.tensor(_A).to(_A) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , _A , atol=_A)) # class_queries_logits __snake_case : List[Any] = outputs.class_queries_logits self.assertEqual(class_queries_logits.shape , (1, model.config.num_queries, model.config.num_labels + 1)) __snake_case : str = torch.tensor( [ [1.8_324, -8.0_835, -4.1_922], [0.8_450, -9.0_050, -3.6_053], [0.3_045, -7.7_293, -3.0_275], ]).to(_A) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , _A , atol=_A)) def _lowercase (self : Any) -> Tuple: __snake_case : List[Any] = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints).to(_A).eval() __snake_case : Union[str, Any] = self.default_image_processor __snake_case : str = image_processor( [np.zeros((3, 8_00, 13_33)), np.zeros((3, 8_00, 13_33))] , segmentation_maps=[np.zeros((3_84, 3_84)).astype(np.floataa), np.zeros((3_84, 3_84)).astype(np.floataa)] , return_tensors='pt' , ) __snake_case : Union[str, Any] = inputs['pixel_values'].to(_A) __snake_case : Optional[Any] = [el.to(_A) for el in inputs['mask_labels']] __snake_case : Tuple = [el.to(_A) for el in inputs['class_labels']] with torch.no_grad(): __snake_case : Optional[int] = model(**_A) self.assertTrue(outputs.loss is not None)
172
"""simple docstring""" import copy import inspect import unittest from transformers import PretrainedConfig, SwiftFormerConfig from transformers.testing_utils import ( require_torch, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwiftFormerForImageClassification, SwiftFormerModel from transformers.models.swiftformer.modeling_swiftformer import SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class UpperCamelCase : def __init__(self : Optional[Any] , _A : Optional[Any] , _A : str=13 , _A : List[Any]=3 , _A : Tuple=True , _A : List[str]=True , _A : Any=0.1 , _A : str=0.1 , _A : Union[str, Any]=2_24 , _A : Dict=10_00 , _A : Optional[int]=[3, 3, 6, 4] , _A : Optional[Any]=[48, 56, 1_12, 2_20] , ) -> List[str]: __snake_case : int = parent __snake_case : str = batch_size __snake_case : int = num_channels __snake_case : Optional[Any] = is_training __snake_case : Tuple = use_labels __snake_case : Optional[Any] = hidden_dropout_prob __snake_case : Optional[int] = attention_probs_dropout_prob __snake_case : Dict = num_labels __snake_case : Union[str, Any] = image_size __snake_case : int = layer_depths __snake_case : List[str] = embed_dims def _lowercase (self : List[Any]) -> Dict: __snake_case : Any = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) __snake_case : int = None if self.use_labels: __snake_case : Tuple = ids_tensor([self.batch_size] , self.num_labels) __snake_case : Optional[int] = self.get_config() return config, pixel_values, labels def _lowercase (self : Union[str, Any]) -> List[Any]: return SwiftFormerConfig( depths=self.layer_depths , embed_dims=self.embed_dims , mlp_ratio=4 , downsamples=[True, True, True, True] , hidden_act='gelu' , num_labels=self.num_labels , down_patch_size=3 , down_stride=2 , down_pad=1 , drop_rate=0.0 , drop_path_rate=0.0 , use_layer_scale=_A , layer_scale_init_value=1E-5 , ) def _lowercase (self : int , _A : Union[str, Any] , _A : Tuple , _A : List[str]) -> Optional[int]: __snake_case : str = SwiftFormerModel(config=_A) model.to(_A) model.eval() __snake_case : Union[str, Any] = model(_A) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dims[-1], 7, 7)) def _lowercase (self : Optional[int] , _A : List[str] , _A : Union[str, Any] , _A : Union[str, Any]) -> int: __snake_case : Optional[int] = self.num_labels __snake_case : Dict = SwiftFormerForImageClassification(_A) model.to(_A) model.eval() __snake_case : List[Any] = model(_A , labels=_A) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) __snake_case : List[str] = SwiftFormerForImageClassification(_A) model.to(_A) model.eval() __snake_case : int = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) __snake_case : Union[str, Any] = model(_A) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def _lowercase (self : Optional[int]) -> int: ((__snake_case) , (__snake_case) , (__snake_case)) : List[Any] = self.prepare_config_and_inputs() __snake_case : str = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class UpperCamelCase ( lowercase , lowercase , unittest.TestCase ): UpperCAmelCase : Union[str, Any] = (SwiftFormerModel, SwiftFormerForImageClassification) if is_torch_available() else () UpperCAmelCase : Union[str, Any] = ( {"""feature-extraction""": SwiftFormerModel, """image-classification""": SwiftFormerForImageClassification} if is_torch_available() else {} ) UpperCAmelCase : Any = False UpperCAmelCase : Any = False UpperCAmelCase : Tuple = False UpperCAmelCase : Tuple = False UpperCAmelCase : Tuple = False def _lowercase (self : int) -> Optional[int]: __snake_case : Dict = SwiftFormerModelTester(self) __snake_case : List[Any] = ConfigTester( self , config_class=_A , has_text_modality=_A , hidden_size=37 , num_attention_heads=12 , num_hidden_layers=12 , ) def _lowercase (self : Dict) -> Dict: self.config_tester.run_common_tests() @unittest.skip(reason='SwiftFormer does not use inputs_embeds') def _lowercase (self : Optional[int]) -> Optional[int]: pass def _lowercase (self : Dict) -> Optional[int]: __snake_case , __snake_case : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : str = model_class(_A) __snake_case : Optional[int] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_A , nn.Linear)) def _lowercase (self : str) -> Any: __snake_case , __snake_case : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : str = model_class(_A) __snake_case : str = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : int = [*signature.parameters.keys()] __snake_case : List[str] = ['pixel_values'] self.assertListEqual(arg_names[:1] , _A) def _lowercase (self : List[Any]) -> List[str]: __snake_case : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_A) def _lowercase (self : Optional[int]) -> int: __snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_A) @slow def _lowercase (self : Union[str, Any]) -> Dict: for model_name in SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : Dict = SwiftFormerModel.from_pretrained(_A) self.assertIsNotNone(_A) @unittest.skip(reason='SwiftFormer does not output attentions') def _lowercase (self : Dict) -> int: pass def _lowercase (self : Union[str, Any]) -> List[Any]: def check_hidden_states_output(_A : str , _A : int , _A : str): __snake_case : Optional[int] = model_class(_A) model.to(_A) model.eval() with torch.no_grad(): __snake_case : Optional[int] = model(**self._prepare_for_class(_A , _A)) __snake_case : Optional[int] = outputs.hidden_states __snake_case : Any = 8 self.assertEqual(len(_A) , _A) # TODO # SwiftFormer's feature maps are of shape (batch_size, embed_dims, height, width) # with the width and height being successively divided by 2, after every 2 blocks for i in range(len(_A)): self.assertEqual( hidden_states[i].shape , torch.Size( [ self.model_tester.batch_size, self.model_tester.embed_dims[i // 2], (self.model_tester.image_size // 4) // 2 ** (i // 2), (self.model_tester.image_size // 4) // 2 ** (i // 2), ]) , ) __snake_case , __snake_case : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Union[str, Any] = True check_hidden_states_output(_A , _A , _A) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : Any = True check_hidden_states_output(_A , _A , _A) def _lowercase (self : List[Any]) -> int: def _config_zero_init(_A : Union[str, Any]): __snake_case : Optional[int] = copy.deepcopy(_A) for key in configs_no_init.__dict__.keys(): if "_range" in key or "_std" in key or "initializer_factor" in key or "layer_scale" in key: setattr(_A , _A , 1E-10) if isinstance(getattr(_A , _A , _A) , _A): __snake_case : Optional[int] = _config_zero_init(getattr(_A , _A)) setattr(_A , _A , _A) return configs_no_init __snake_case , __snake_case : Tuple = self.model_tester.prepare_config_and_inputs_for_common() __snake_case : int = _config_zero_init(_A) for model_class in self.all_model_classes: __snake_case : Tuple = model_class(config=_A) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1E9) / 1E9).round().item() , [0.0, 1.0] , msg=f"Parameter {name} of model {model_class} seems not properly initialized" , ) @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.') def _lowercase (self : List[str]) -> List[Any]: pass def __UpperCAmelCase ( ) -> List[str]: '''simple docstring''' __snake_case : Optional[int] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class UpperCamelCase ( unittest.TestCase ): @cached_property def _lowercase (self : Dict) -> List[str]: return ViTImageProcessor.from_pretrained('MBZUAI/swiftformer-xs') if is_vision_available() else None @slow def _lowercase (self : Any) -> List[Any]: __snake_case : Any = SwiftFormerForImageClassification.from_pretrained('MBZUAI/swiftformer-xs').to(_A) __snake_case : Any = self.default_image_processor __snake_case : Optional[int] = prepare_img() __snake_case : Optional[int] = image_processor(images=_A , return_tensors='pt').to(_A) # forward pass with torch.no_grad(): __snake_case : Tuple = model(**_A) # verify the logits __snake_case : Optional[Any] = torch.Size((1, 10_00)) self.assertEqual(outputs.logits.shape , _A) __snake_case : Optional[int] = torch.tensor([[-2.1_703E00, 2.1_107E00, -2.0_811E00]]).to(_A) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _A , atol=1E-4))
172
1
# Copyright 2022 The HuggingFace Team and The OpenBMB Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __a = { 'configuration_cpmant': ['CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'CpmAntConfig'], 'tokenization_cpmant': ['CpmAntTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ 'CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST', 'CpmAntForCausalLM', 'CpmAntModel', 'CpmAntPreTrainedModel', ] if TYPE_CHECKING: from .configuration_cpmant import CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP, CpmAntConfig from .tokenization_cpmant import CpmAntTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_cpmant import ( CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST, CpmAntForCausalLM, CpmAntModel, CpmAntPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
362
import tempfile import torch from diffusers import IPNDMScheduler from .test_schedulers import SchedulerCommonTest class __a( _a ): """simple docstring""" lowerCAmelCase = (IPNDMScheduler,) lowerCAmelCase = (('''num_inference_steps''', 50),) def a__ ( self ,**_SCREAMING_SNAKE_CASE ) -> int: UpperCAmelCase_ : Tuple = {'''num_train_timesteps''': 1_000} config.update(**_SCREAMING_SNAKE_CASE ) return config def a__ ( self ,_SCREAMING_SNAKE_CASE=0 ,**_SCREAMING_SNAKE_CASE ) -> Union[str, Any]: UpperCAmelCase_ : Optional[int] = dict(self.forward_default_kwargs ) UpperCAmelCase_ : int = kwargs.pop('''num_inference_steps''' ,_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : str = self.dummy_sample UpperCAmelCase_ : Optional[Any] = 0.1 * sample UpperCAmelCase_ : Optional[int] = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : Tuple = self.get_scheduler_config(**_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Optional[int] = scheduler_class(**_SCREAMING_SNAKE_CASE ) scheduler.set_timesteps(_SCREAMING_SNAKE_CASE ) # copy over dummy past residuals UpperCAmelCase_ : Optional[Any] = dummy_past_residuals[:] if time_step is None: UpperCAmelCase_ : Tuple = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : List[Any] = scheduler_class.from_pretrained(_SCREAMING_SNAKE_CASE ) new_scheduler.set_timesteps(_SCREAMING_SNAKE_CASE ) # copy over dummy past residuals UpperCAmelCase_ : Optional[int] = dummy_past_residuals[:] UpperCAmelCase_ : List[Any] = scheduler.step(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ).prev_sample UpperCAmelCase_ : str = new_scheduler.step(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" UpperCAmelCase_ : Union[str, Any] = scheduler.step(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ).prev_sample UpperCAmelCase_ : Optional[int] = new_scheduler.step(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def a__ ( self ) -> List[Any]: pass def a__ ( self ,_SCREAMING_SNAKE_CASE=0 ,**_SCREAMING_SNAKE_CASE ) -> List[Any]: UpperCAmelCase_ : List[str] = dict(self.forward_default_kwargs ) UpperCAmelCase_ : Tuple = kwargs.pop('''num_inference_steps''' ,_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : int = self.dummy_sample UpperCAmelCase_ : Tuple = 0.1 * sample UpperCAmelCase_ : int = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : str = self.get_scheduler_config() UpperCAmelCase_ : str = scheduler_class(**_SCREAMING_SNAKE_CASE ) scheduler.set_timesteps(_SCREAMING_SNAKE_CASE ) # copy over dummy past residuals (must be after setting timesteps) UpperCAmelCase_ : Tuple = dummy_past_residuals[:] if time_step is None: UpperCAmelCase_ : Dict = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Optional[int] = scheduler_class.from_pretrained(_SCREAMING_SNAKE_CASE ) # copy over dummy past residuals new_scheduler.set_timesteps(_SCREAMING_SNAKE_CASE ) # copy over dummy past residual (must be after setting timesteps) UpperCAmelCase_ : str = dummy_past_residuals[:] UpperCAmelCase_ : Tuple = scheduler.step(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ).prev_sample UpperCAmelCase_ : Optional[int] = new_scheduler.step(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" UpperCAmelCase_ : str = scheduler.step(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ).prev_sample UpperCAmelCase_ : List[str] = new_scheduler.step(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def a__ ( self ,**_SCREAMING_SNAKE_CASE ) -> List[Any]: UpperCAmelCase_ : Tuple = self.scheduler_classes[0] UpperCAmelCase_ : List[Any] = self.get_scheduler_config(**_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Tuple = scheduler_class(**_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Any = 10 UpperCAmelCase_ : Tuple = self.dummy_model() UpperCAmelCase_ : List[str] = self.dummy_sample_deter scheduler.set_timesteps(_SCREAMING_SNAKE_CASE ) for i, t in enumerate(scheduler.timesteps ): UpperCAmelCase_ : int = model(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Union[str, Any] = scheduler.step(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ).prev_sample for i, t in enumerate(scheduler.timesteps ): UpperCAmelCase_ : Optional[int] = model(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Dict = scheduler.step(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ).prev_sample return sample def a__ ( self ) -> str: UpperCAmelCase_ : List[Any] = dict(self.forward_default_kwargs ) UpperCAmelCase_ : Any = kwargs.pop('''num_inference_steps''' ,_SCREAMING_SNAKE_CASE ) for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : int = self.get_scheduler_config() UpperCAmelCase_ : List[Any] = scheduler_class(**_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Optional[int] = self.dummy_sample UpperCAmelCase_ : List[Any] = 0.1 * sample if num_inference_steps is not None and hasattr(_SCREAMING_SNAKE_CASE ,'''set_timesteps''' ): scheduler.set_timesteps(_SCREAMING_SNAKE_CASE ) elif num_inference_steps is not None and not hasattr(_SCREAMING_SNAKE_CASE ,'''set_timesteps''' ): UpperCAmelCase_ : Union[str, Any] = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) UpperCAmelCase_ : str = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] UpperCAmelCase_ : int = dummy_past_residuals[:] UpperCAmelCase_ : List[Any] = scheduler.timesteps[5] UpperCAmelCase_ : Dict = scheduler.timesteps[6] UpperCAmelCase_ : List[Any] = scheduler.step(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ).prev_sample UpperCAmelCase_ : Dict = scheduler.step(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ).prev_sample self.assertEqual(output_a.shape ,sample.shape ) self.assertEqual(output_a.shape ,output_a.shape ) UpperCAmelCase_ : List[Any] = scheduler.step(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ).prev_sample UpperCAmelCase_ : Dict = scheduler.step(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ).prev_sample self.assertEqual(output_a.shape ,sample.shape ) self.assertEqual(output_a.shape ,output_a.shape ) def a__ ( self ) -> Any: for timesteps in [100, 1_000]: self.check_over_configs(num_train_timesteps=_SCREAMING_SNAKE_CASE ,time_step=_SCREAMING_SNAKE_CASE ) def a__ ( self ) -> Union[str, Any]: for t, num_inference_steps in zip([1, 5, 10] ,[10, 50, 100] ): self.check_over_forward(num_inference_steps=_SCREAMING_SNAKE_CASE ,time_step=_SCREAMING_SNAKE_CASE ) def a__ ( self ) -> Union[str, Any]: UpperCAmelCase_ : Optional[int] = self.full_loop() UpperCAmelCase_ : str = torch.mean(torch.abs(_SCREAMING_SNAKE_CASE ) ) assert abs(result_mean.item() - 2_540_529 ) < 10
235
0
import logging import os import sys from pathlib import Path from unittest.mock import patch from parameterized import parameterized from run_eval import run_generate from run_eval_search import run_search from transformers.testing_utils import CaptureStdout, TestCasePlus, slow from utils import ROUGE_KEYS logging.basicConfig(level=logging.DEBUG) _UpperCAmelCase : List[str] = logging.getLogger() def A ( lowercase , lowercase ) -> Optional[int]: '''simple docstring''' UpperCamelCase = '\n'.join(lowercase ) Path(lowercase ).open('w' ).writelines(lowercase ) _UpperCAmelCase : Union[str, Any] = "patrickvonplaten/t5-tiny-random" _UpperCAmelCase : Optional[int] = "sshleifer/bart-tiny-random" _UpperCAmelCase : List[Any] = "sshleifer/tiny-mbart" _UpperCAmelCase : Optional[Any] = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) logging.disable(logging.CRITICAL) # remove noisy download output from tracebacks class lowercase ( _SCREAMING_SNAKE_CASE ): def __UpperCamelCase ( self , A_ ) -> List[str]: """simple docstring""" UpperCamelCase = Path(self.get_auto_remove_tmp_dir() ) / 'utest_input.source' UpperCamelCase = input_file_name.parent / 'utest_output.txt' assert not output_file_name.exists() UpperCamelCase = [' New York (CNN)When Liana Barrientos was 23 years old, she got married in Westchester County.'] _dump_articles(A_ , A_ ) UpperCamelCase = str(Path(self.get_auto_remove_tmp_dir() ) / 'scores.json' ) UpperCamelCase = 'translation_en_to_de' if model == T5_TINY else 'summarization' UpperCamelCase = F''' run_eval_search.py {model} {input_file_name} {output_file_name} --score_path {score_path} --task {task} --num_beams 2 --length_penalty 2.0 '''.split() with patch.object(A_ , 'argv' , A_ ): run_generate() assert Path(A_ ).exists() # os.remove(Path(output_file_name)) def __UpperCamelCase ( self ) -> List[str]: """simple docstring""" self.run_eval_tester(A_ ) @parameterized.expand([BART_TINY, MBART_TINY] ) @slow def __UpperCamelCase ( self , A_ ) -> Dict: """simple docstring""" self.run_eval_tester(A_ ) @parameterized.expand([T5_TINY, MBART_TINY] ) @slow def __UpperCamelCase ( self , A_ ) -> Tuple: """simple docstring""" UpperCamelCase = Path(self.get_auto_remove_tmp_dir() ) / 'utest_input.source' UpperCamelCase = input_file_name.parent / 'utest_output.txt' assert not output_file_name.exists() UpperCamelCase = { 'en': ['Machine learning is great, isn\'t it?', 'I like to eat bananas', 'Tomorrow is another great day!'], 'de': [ 'Maschinelles Lernen ist großartig, oder?', 'Ich esse gerne Bananen', 'Morgen ist wieder ein toller Tag!', ], } UpperCamelCase = Path(self.get_auto_remove_tmp_dir() ) UpperCamelCase = str(tmp_dir / 'scores.json' ) UpperCamelCase = str(tmp_dir / 'val.target' ) _dump_articles(A_ , text['en'] ) _dump_articles(A_ , text['de'] ) UpperCamelCase = 'translation_en_to_de' if model == T5_TINY else 'summarization' UpperCamelCase = F''' run_eval_search.py {model} {str(A_ )} {str(A_ )} --score_path {score_path} --reference_path {reference_path} --task {task} '''.split() testargs.extend(['--search', 'num_beams=1:2 length_penalty=0.9:1.0'] ) with patch.object(A_ , 'argv' , A_ ): with CaptureStdout() as cs: run_search() UpperCamelCase = [' num_beams | length_penalty', model, 'Best score args'] UpperCamelCase = ['Info'] if "translation" in task: expected_strings.append('bleu' ) else: expected_strings.extend(A_ ) for w in expected_strings: assert w in cs.out for w in un_expected_strings: assert w not in cs.out assert Path(A_ ).exists() os.remove(Path(A_ ) )
222
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _UpperCAmelCase : List[Any] = logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = { "Helsinki-NLP/opus-mt-en-de": "https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json", # See all Marian models at https://huggingface.co/models?filter=marian } class lowercase ( _SCREAMING_SNAKE_CASE ): __lowercase : int = "marian" __lowercase : List[str] = ["past_key_values"] __lowercase : Any = {"num_attention_heads": "encoder_attention_heads", "hidden_size": "d_model"} def __init__( self , A_=58_101 , A_=None , A_=1_024 , A_=12 , A_=4_096 , A_=16 , A_=12 , A_=4_096 , A_=16 , A_=0.0 , A_=0.0 , A_=True , A_=True , A_="gelu" , A_=1_024 , A_=0.1 , A_=0.0 , A_=0.0 , A_=0.02 , A_=58_100 , A_=False , A_=58_100 , A_=0 , A_=0 , A_=True , **A_ , ) -> Optional[int]: """simple docstring""" UpperCamelCase = vocab_size UpperCamelCase = decoder_vocab_size or vocab_size UpperCamelCase = max_position_embeddings UpperCamelCase = d_model UpperCamelCase = encoder_ffn_dim UpperCamelCase = encoder_layers UpperCamelCase = encoder_attention_heads UpperCamelCase = decoder_ffn_dim UpperCamelCase = decoder_layers UpperCamelCase = decoder_attention_heads UpperCamelCase = dropout UpperCamelCase = attention_dropout UpperCamelCase = activation_dropout UpperCamelCase = activation_function UpperCamelCase = init_std UpperCamelCase = encoder_layerdrop UpperCamelCase = decoder_layerdrop UpperCamelCase = use_cache UpperCamelCase = encoder_layers UpperCamelCase = scale_embedding # scale factor will be sqrt(d_model) if True UpperCamelCase = share_encoder_decoder_embeddings super().__init__( pad_token_id=A_ , eos_token_id=A_ , is_encoder_decoder=A_ , decoder_start_token_id=A_ , forced_eos_token_id=A_ , **A_ , ) class lowercase ( _SCREAMING_SNAKE_CASE ): @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def __UpperCamelCase ( self ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task in ["default", "seq2seq-lm"]: UpperCamelCase = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ] ) if self.use_past: UpperCamelCase = {0: 'batch'} UpperCamelCase = {0: 'batch', 1: 'past_decoder_sequence + sequence'} else: UpperCamelCase = {0: 'batch', 1: 'decoder_sequence'} UpperCamelCase = {0: 'batch', 1: 'decoder_sequence'} if self.use_past: self.fill_with_past_key_values_(A_ , direction='inputs' ) elif self.task == "causal-lm": # TODO: figure this case out. UpperCamelCase = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ] ) if self.use_past: UpperCamelCase , UpperCamelCase = self.num_layers for i in range(A_ ): UpperCamelCase = {0: 'batch', 2: 'past_sequence + sequence'} UpperCamelCase = {0: 'batch', 2: 'past_sequence + sequence'} else: UpperCamelCase = OrderedDict( [ ('input_ids', {0: 'batch', 1: 'encoder_sequence'}), ('attention_mask', {0: 'batch', 1: 'encoder_sequence'}), ('decoder_input_ids', {0: 'batch', 1: 'decoder_sequence'}), ('decoder_attention_mask', {0: 'batch', 1: 'decoder_sequence'}), ] ) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def __UpperCamelCase ( self ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task in ["default", "seq2seq-lm"]: UpperCamelCase = super().outputs else: UpperCamelCase = super(A_ , self ).outputs if self.use_past: UpperCamelCase , UpperCamelCase = self.num_layers for i in range(A_ ): UpperCamelCase = {0: 'batch', 2: 'past_sequence + sequence'} UpperCamelCase = {0: 'batch', 2: 'past_sequence + sequence'} return common_outputs def __UpperCamelCase ( self , A_ , A_ = -1 , A_ = -1 , A_ = False , A_ = None , ) -> Mapping[str, Any]: """simple docstring""" UpperCamelCase = self._generate_dummy_inputs_for_encoder_and_decoder( A_ , A_ , A_ , A_ , A_ ) # Generate decoder inputs UpperCamelCase = seq_length if not self.use_past else 1 UpperCamelCase = self._generate_dummy_inputs_for_encoder_and_decoder( A_ , A_ , A_ , A_ , A_ ) UpperCamelCase = {F'''decoder_{name}''': tensor for name, tensor in decoder_inputs.items()} UpperCamelCase = dict(**A_ , **A_ ) if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch UpperCamelCase , UpperCamelCase = common_inputs['input_ids'].shape UpperCamelCase = common_inputs['decoder_input_ids'].shape[1] UpperCamelCase , UpperCamelCase = self.num_attention_heads UpperCamelCase = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) UpperCamelCase = decoder_seq_length + 3 UpperCamelCase = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) UpperCamelCase = torch.cat( [common_inputs['decoder_attention_mask'], torch.ones(A_ , A_ )] , dim=1 ) UpperCamelCase = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered UpperCamelCase , UpperCamelCase = self.num_layers UpperCamelCase = min(A_ , A_ ) UpperCamelCase = max(A_ , A_ ) - min_num_layers UpperCamelCase = 'encoder' if num_encoder_layers > num_decoder_layers else 'decoder' for _ in range(A_ ): common_inputs["past_key_values"].append( ( torch.zeros(A_ ), torch.zeros(A_ ), torch.zeros(A_ ), torch.zeros(A_ ), ) ) # TODO: test this. UpperCamelCase = encoder_shape if remaining_side_name == 'encoder' else decoder_shape for _ in range(A_ , A_ ): common_inputs["past_key_values"].append((torch.zeros(A_ ), torch.zeros(A_ )) ) return common_inputs def __UpperCamelCase ( self , A_ , A_ = -1 , A_ = -1 , A_ = False , A_ = None , ) -> Mapping[str, Any]: """simple docstring""" UpperCamelCase = self._generate_dummy_inputs_for_encoder_and_decoder( A_ , A_ , A_ , A_ , A_ ) if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch UpperCamelCase , UpperCamelCase = common_inputs['input_ids'].shape # Not using the same length for past_key_values UpperCamelCase = seqlen + 2 UpperCamelCase , UpperCamelCase = self.num_layers UpperCamelCase , UpperCamelCase = self.num_attention_heads UpperCamelCase = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) UpperCamelCase = common_inputs['attention_mask'].dtype UpperCamelCase = torch.cat( [common_inputs['attention_mask'], torch.ones(A_ , A_ , dtype=A_ )] , dim=1 ) UpperCamelCase = [ (torch.zeros(A_ ), torch.zeros(A_ )) for _ in range(A_ ) ] return common_inputs def __UpperCamelCase ( self , A_ , A_ = -1 , A_ = -1 , A_ = False , A_ = None , ) -> Mapping[str, Any]: """simple docstring""" # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX UpperCamelCase = compute_effective_axis_dimension( A_ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX UpperCamelCase = tokenizer.num_special_tokens_to_add(A_ ) UpperCamelCase = compute_effective_axis_dimension( A_ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=A_ ) # Generate dummy inputs according to compute batch and sequence UpperCamelCase = [' '.join([tokenizer.unk_token] ) * seq_length] * batch_size UpperCamelCase = dict(tokenizer(A_ , return_tensors=A_ ) ) return common_inputs def __UpperCamelCase ( self , A_ , A_ = -1 , A_ = -1 , A_ = False , A_ = None , ) -> Mapping[str, Any]: """simple docstring""" if self.task in ["default", "seq2seq-lm"]: UpperCamelCase = self._generate_dummy_inputs_for_default_and_seqaseq_lm( A_ , batch_size=A_ , seq_length=A_ , is_pair=A_ , framework=A_ ) else: UpperCamelCase = self._generate_dummy_inputs_for_causal_lm( A_ , batch_size=A_ , seq_length=A_ , is_pair=A_ , framework=A_ ) return common_inputs def __UpperCamelCase ( self , A_ , A_ , A_ , A_ ) -> int: """simple docstring""" if self.task in ["default", "seq2seq-lm"]: UpperCamelCase = super()._flatten_past_key_values_(A_ , A_ , A_ , A_ ) else: UpperCamelCase = super(A_ , self )._flatten_past_key_values_( A_ , A_ , A_ , A_ ) @property def __UpperCamelCase ( self ) -> float: """simple docstring""" return 1e-4
222
1
import warnings from ...utils import logging from .image_processing_perceiver import PerceiverImageProcessor _SCREAMING_SNAKE_CASE : Tuple = logging.get_logger(__name__) class A__ ( snake_case__ ): """simple docstring""" def __init__( self , *__snake_case , **__snake_case ): warnings.warn( '''The class PerceiverFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use PerceiverImageProcessor instead.''' , __snake_case , ) super().__init__(*__snake_case , **__snake_case )
213
# tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. _SCREAMING_SNAKE_CASE : Union[str, Any] = abspath(join(dirname(dirname(dirname(__file__))), "src")) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action="ignore", category=FutureWarning) def UpperCAmelCase__ (UpperCamelCase_ ): """simple docstring""" from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(UpperCamelCase_ ) def UpperCAmelCase__ (UpperCamelCase_ ): """simple docstring""" from transformers.testing_utils import pytest_terminal_summary_main snake_case = terminalreporter.config.getoption('''--make-reports''' ) if make_reports: pytest_terminal_summary_main(UpperCamelCase_ ,id=UpperCamelCase_ )
213
1
from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import ScoreSdeVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class SCREAMING_SNAKE_CASE__ ( lowercase__ ): snake_case__ : UNetaDModel snake_case__ : ScoreSdeVeScheduler def __init__( self : str , SCREAMING_SNAKE_CASE__ : UNetaDModel , SCREAMING_SNAKE_CASE__ : ScoreSdeVeScheduler ) -> Optional[int]: super().__init__() self.register_modules(unet=SCREAMING_SNAKE_CASE__ , scheduler=SCREAMING_SNAKE_CASE__ ) @torch.no_grad() def __call__( self : str , SCREAMING_SNAKE_CASE__ : int = 1 , SCREAMING_SNAKE_CASE__ : int = 2_0_0_0 , SCREAMING_SNAKE_CASE__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , SCREAMING_SNAKE_CASE__ : Optional[str] = "pil" , SCREAMING_SNAKE_CASE__ : bool = True , **SCREAMING_SNAKE_CASE__ : str , ) -> Union[ImagePipelineOutput, Tuple]: a_ : Tuple = self.unet.config.sample_size a_ : Optional[Any] = (batch_size, 3, img_size, img_size) a_ : Optional[Any] = self.unet a_ : List[str] = randn_tensor(SCREAMING_SNAKE_CASE__ , generator=SCREAMING_SNAKE_CASE__ ) * self.scheduler.init_noise_sigma a_ : Tuple = sample.to(self.device ) self.scheduler.set_timesteps(SCREAMING_SNAKE_CASE__ ) self.scheduler.set_sigmas(SCREAMING_SNAKE_CASE__ ) for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): a_ : Dict = self.scheduler.sigmas[i] * torch.ones(shape[0] , device=self.device ) # correction step for _ in range(self.scheduler.config.correct_steps ): a_ : Tuple = self.unet(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ).sample a_ : Optional[Any] = self.scheduler.step_correct(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , generator=SCREAMING_SNAKE_CASE__ ).prev_sample # prediction step a_ : Optional[int] = model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ).sample a_ : Tuple = self.scheduler.step_pred(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , generator=SCREAMING_SNAKE_CASE__ ) a_ , a_ : Union[str, Any] = output.prev_sample, output.prev_sample_mean a_ : Dict = sample_mean.clamp(0 , 1 ) a_ : Optional[Any] = sample.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": a_ : List[Any] = self.numpy_to_pil(SCREAMING_SNAKE_CASE__ ) if not return_dict: return (sample,) return ImagePipelineOutput(images=SCREAMING_SNAKE_CASE__ )
32
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available A__: List[str] = { '''configuration_chinese_clip''': [ '''CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ChineseCLIPConfig''', '''ChineseCLIPOnnxConfig''', '''ChineseCLIPTextConfig''', '''ChineseCLIPVisionConfig''', ], '''processing_chinese_clip''': ['''ChineseCLIPProcessor'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__: Optional[int] = ['''ChineseCLIPFeatureExtractor'''] A__: Any = ['''ChineseCLIPImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__: Dict = [ '''CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ChineseCLIPModel''', '''ChineseCLIPPreTrainedModel''', '''ChineseCLIPTextModel''', '''ChineseCLIPVisionModel''', ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys A__: str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
276
0
from __future__ import annotations from typing import Any class lowercase_ : def __init__( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = 0 ): """simple docstring""" UpperCamelCase_ , UpperCamelCase_ = row, column UpperCamelCase_ = [[default_value for c in range(__UpperCamelCase )] for r in range(__UpperCamelCase )] def __str__( self ): """simple docstring""" UpperCamelCase_ = f'''Matrix consist of {self.row} rows and {self.column} columns\n''' # Make string identifier UpperCamelCase_ = 0 for row_vector in self.array: for obj in row_vector: UpperCamelCase_ = max(__UpperCamelCase , len(str(__UpperCamelCase ) ) ) UpperCamelCase_ = f'''%{max_element_length}s''' # Make string and return def single_line(__UpperCamelCase ) -> str: nonlocal string_format_identifier UpperCamelCase_ = """[""" line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(__UpperCamelCase ) for row_vector in self.array ) return s def __repr__( self ): """simple docstring""" return str(self ) def lowerCamelCase_ ( self , __UpperCamelCase ): """simple docstring""" if not (isinstance(__UpperCamelCase , (list, tuple) ) and len(__UpperCamelCase ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self , __UpperCamelCase ): """simple docstring""" assert self.validate_indicies(__UpperCamelCase ) return self.array[loc[0]][loc[1]] def __setitem__( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" assert self.validate_indicies(__UpperCamelCase ) UpperCamelCase_ = value def __add__( self , __UpperCamelCase ): """simple docstring""" assert isinstance(__UpperCamelCase , __UpperCamelCase ) assert self.row == another.row and self.column == another.column # Add UpperCamelCase_ = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): UpperCamelCase_ = self[r, c] + another[r, c] return result def __neg__( self ): """simple docstring""" UpperCamelCase_ = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): UpperCamelCase_ = -self[r, c] return result def __sub__( self , __UpperCamelCase ): """simple docstring""" return self + (-another) def __mul__( self , __UpperCamelCase ): """simple docstring""" if isinstance(__UpperCamelCase , (int, float) ): # Scalar multiplication UpperCamelCase_ = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): UpperCamelCase_ = self[r, c] * another return result elif isinstance(__UpperCamelCase , __UpperCamelCase ): # Matrix multiplication assert self.column == another.row UpperCamelCase_ = Matrix(self.row , another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: UpperCamelCase_ = f'''Unsupported type given for another ({type(__UpperCamelCase )})''' raise TypeError(__UpperCamelCase ) def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = Matrix(self.column , self.row ) for r in range(self.row ): for c in range(self.column ): UpperCamelCase_ = self[r, c] return result def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" assert isinstance(__UpperCamelCase , __UpperCamelCase ) and isinstance(__UpperCamelCase , __UpperCamelCase ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate UpperCamelCase_ = v.transpose() UpperCamelCase_ = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def lowerCamelCase__ ( ) -> None: # a^(-1) UpperCamelCase_ = Matrix(3 , 3 , 0 ) for i in range(3 ): UpperCamelCase_ = 1 print(f'''a^(-1) is {ainv}''' ) # u, v UpperCamelCase_ = Matrix(3 , 1 , 0 ) UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ = 1, 2, -3 UpperCamelCase_ = Matrix(3 , 1 , 0 ) UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ = 4, -2, 5 print(f'''u is {u}''' ) print(f'''v is {v}''' ) print(f'''uv^T is {u * v.transpose()}''' ) # Sherman Morrison print(f'''(a + uv^T)^(-1) is {ainv.sherman_morrison(a__ , a__ )}''' ) def lowerCamelCase__ ( ) -> None: import doctest doctest.testmod() testa()
359
import copy import re class lowercase_ : A__ : Optional[Any] = """hp""" A__ : Union[str, Any] = {} A__ : Optional[int] = None @classmethod def lowerCamelCase_ ( cls , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" UpperCamelCase_ = prefix UpperCamelCase_ = defaults cls.build_naming_info() @staticmethod def lowerCamelCase_ ( __UpperCamelCase , __UpperCamelCase ): """simple docstring""" if len(__UpperCamelCase ) == 0: return "" UpperCamelCase_ = None if any(char.isdigit() for char in word ): raise Exception(f'''Parameters should not contain numbers: \'{word}\' contains a number''' ) if word in info["short_word"]: return info["short_word"][word] for prefix_len in range(1 , len(__UpperCamelCase ) + 1 ): UpperCamelCase_ = word[:prefix_len] if prefix in info["reverse_short_word"]: continue else: UpperCamelCase_ = prefix break if short_word is None: # Paranoid fallback def int_to_alphabetic(__UpperCamelCase ): UpperCamelCase_ = """""" while integer != 0: UpperCamelCase_ = chr(ord("""A""" ) + integer % 1_0 ) + s integer //= 1_0 return s UpperCamelCase_ = 0 while True: UpperCamelCase_ = word + """#""" + int_to_alphabetic(__UpperCamelCase ) if sword in info["reverse_short_word"]: continue else: UpperCamelCase_ = sword break UpperCamelCase_ = short_word UpperCamelCase_ = word return short_word @staticmethod def lowerCamelCase_ ( __UpperCamelCase , __UpperCamelCase ): """simple docstring""" UpperCamelCase_ = param_name.split("""_""" ) UpperCamelCase_ = [TrialShortNamer.shortname_for_word(__UpperCamelCase , __UpperCamelCase ) for word in words] # We try to create a separatorless short name, but if there is a collision we have to fallback # to a separated short name UpperCamelCase_ = ["""""", """_"""] for separator in separators: UpperCamelCase_ = separator.join(__UpperCamelCase ) if shortname not in info["reverse_short_param"]: UpperCamelCase_ = shortname UpperCamelCase_ = param_name return shortname return param_name @staticmethod def lowerCamelCase_ ( __UpperCamelCase , __UpperCamelCase ): """simple docstring""" UpperCamelCase_ = TrialShortNamer.shortname_for_key(__UpperCamelCase , __UpperCamelCase ) UpperCamelCase_ = short_name UpperCamelCase_ = param_name @classmethod def lowerCamelCase_ ( cls ): """simple docstring""" if cls.NAMING_INFO is not None: return UpperCamelCase_ = { """short_word""": {}, """reverse_short_word""": {}, """short_param""": {}, """reverse_short_param""": {}, } UpperCamelCase_ = list(cls.DEFAULTS.keys() ) for k in field_keys: cls.add_new_param_name(__UpperCamelCase , __UpperCamelCase ) UpperCamelCase_ = info @classmethod def lowerCamelCase_ ( cls , __UpperCamelCase ): """simple docstring""" cls.build_naming_info() assert cls.PREFIX is not None UpperCamelCase_ = [copy.copy(cls.PREFIX )] for k, v in params.items(): if k not in cls.DEFAULTS: raise Exception(f'''You should provide a default value for the param name {k} with value {v}''' ) if v == cls.DEFAULTS[k]: # The default value is not added to the name continue UpperCamelCase_ = cls.NAMING_INFO["""short_param"""][k] if isinstance(__UpperCamelCase , __UpperCamelCase ): UpperCamelCase_ = 1 if v else 0 UpperCamelCase_ = """""" if isinstance(__UpperCamelCase , (int, float) ) else """-""" UpperCamelCase_ = f'''{key}{sep}{v}''' name.append(__UpperCamelCase ) return "_".join(__UpperCamelCase ) @classmethod def lowerCamelCase_ ( cls , __UpperCamelCase ): """simple docstring""" UpperCamelCase_ = repr[len(cls.PREFIX ) + 1 :] if repr == "": UpperCamelCase_ = [] else: UpperCamelCase_ = repr.split("""_""" ) UpperCamelCase_ = {} for value in values: if "-" in value: UpperCamelCase_ , UpperCamelCase_ = value.split("""-""" ) else: UpperCamelCase_ = re.sub("""[0-9.]""" , """""" , __UpperCamelCase ) UpperCamelCase_ = float(re.sub("""[^0-9.]""" , """""" , __UpperCamelCase ) ) UpperCamelCase_ = cls.NAMING_INFO["""reverse_short_param"""][p_k] UpperCamelCase_ = p_v for k in cls.DEFAULTS: if k not in parameters: UpperCamelCase_ = cls.DEFAULTS[k] return parameters
261
0
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL import torch from transformers import CLIPImageProcessor, CLIPVisionModel from ...models import PriorTransformer from ...pipelines import DiffusionPipeline from ...schedulers import HeunDiscreteScheduler from ...utils import ( BaseOutput, is_accelerate_available, logging, randn_tensor, replace_example_docstring, ) from .renderer import ShapERenderer __A =logging.get_logger(__name__) # pylint: disable=invalid-name __A =''' Examples: ```py >>> from PIL import Image >>> import torch >>> from diffusers import DiffusionPipeline >>> from diffusers.utils import export_to_gif, load_image >>> device = torch.device("cuda" if torch.cuda.is_available() else "cpu") >>> repo = "openai/shap-e-img2img" >>> pipe = DiffusionPipeline.from_pretrained(repo, torch_dtype=torch.float16) >>> pipe = pipe.to(device) >>> guidance_scale = 3.0 >>> image_url = "https://hf.co/datasets/diffusers/docs-images/resolve/main/shap-e/corgi.png" >>> image = load_image(image_url).convert("RGB") >>> images = pipe( ... image, ... guidance_scale=guidance_scale, ... num_inference_steps=64, ... frame_size=256, ... ).images >>> gif_path = export_to_gif(images[0], "corgi_3d.gif") ``` ''' @dataclass class _SCREAMING_SNAKE_CASE ( snake_case_ ): lowerCAmelCase__ = 42 class _SCREAMING_SNAKE_CASE ( snake_case_ ): def __init__( self , lowercase , lowercase , lowercase , lowercase , lowercase , ) -> List[str]: super().__init__() self.register_modules( prior=lowercase , image_encoder=lowercase , image_processor=lowercase , scheduler=lowercase , renderer=lowercase , ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ) -> int: if latents is None: lowerCamelCase_ = randn_tensor(lowercase , generator=lowercase , device=lowercase , dtype=lowercase ) else: if latents.shape != shape: raise ValueError(f'Unexpected latents shape, got {latents.shape}, expected {shape}' ) lowerCamelCase_ = latents.to(lowercase ) lowerCamelCase_ = latents * scheduler.init_noise_sigma return latents def SCREAMING_SNAKE_CASE_( self , lowercase=0 ) -> int: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) lowerCamelCase_ = torch.device(f'cuda:{gpu_id}' ) lowerCamelCase_ = [self.image_encoder, self.prior] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(lowercase , lowercase ) @property def SCREAMING_SNAKE_CASE_( self ) -> List[str]: if self.device != torch.device("meta" ) or not hasattr(self.image_encoder , "_hf_hook" ): return self.device for module in self.image_encoder.modules(): if ( hasattr(lowercase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , ) -> List[str]: if isinstance(lowercase , lowercase ) and isinstance(image[0] , torch.Tensor ): lowerCamelCase_ = torch.cat(lowercase , axis=0 ) if image[0].ndim == 4 else torch.stack(lowercase , axis=0 ) if not isinstance(lowercase , torch.Tensor ): lowerCamelCase_ = self.image_processor(lowercase , return_tensors="pt" ).pixel_values[0].unsqueeze(0 ) lowerCamelCase_ = image.to(dtype=self.image_encoder.dtype , device=lowercase ) lowerCamelCase_ = self.image_encoder(lowercase )["last_hidden_state"] lowerCamelCase_ = image_embeds[:, 1:, :].contiguous() # batch_size, dim, 256 lowerCamelCase_ = image_embeds.repeat_interleave(lowercase , dim=0 ) if do_classifier_free_guidance: lowerCamelCase_ = torch.zeros_like(lowercase ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes lowerCamelCase_ = torch.cat([negative_image_embeds, image_embeds] ) return image_embeds @torch.no_grad() @replace_example_docstring(lowercase ) def __call__( self , lowercase , lowercase = 1 , lowercase = 25 , lowercase = None , lowercase = None , lowercase = 4.0 , lowercase = 64 , lowercase = "pil" , lowercase = True , ) -> Union[str, Any]: if isinstance(lowercase , PIL.Image.Image ): lowerCamelCase_ = 1 elif isinstance(lowercase , torch.Tensor ): lowerCamelCase_ = image.shape[0] elif isinstance(lowercase , lowercase ) and isinstance(image[0] , (torch.Tensor, PIL.Image.Image) ): lowerCamelCase_ = len(lowercase ) else: raise ValueError( f'`image` has to be of type `PIL.Image.Image`, `torch.Tensor`, `List[PIL.Image.Image]` or `List[torch.Tensor]` but is {type(lowercase )}' ) lowerCamelCase_ = self._execution_device lowerCamelCase_ = batch_size * num_images_per_prompt lowerCamelCase_ = guidance_scale > 1.0 lowerCamelCase_ = self._encode_image(lowercase , lowercase , lowercase , lowercase ) # prior self.scheduler.set_timesteps(lowercase , device=lowercase ) lowerCamelCase_ = self.scheduler.timesteps lowerCamelCase_ = self.prior.config.num_embeddings lowerCamelCase_ = self.prior.config.embedding_dim lowerCamelCase_ = self.prepare_latents( (batch_size, num_embeddings * embedding_dim) , image_embeds.dtype , lowercase , lowercase , lowercase , self.scheduler , ) # YiYi notes: for testing only to match ldm, we can directly create a latents with desired shape: batch_size, num_embeddings, embedding_dim lowerCamelCase_ = latents.reshape(latents.shape[0] , lowercase , lowercase ) for i, t in enumerate(self.progress_bar(lowercase ) ): # expand the latents if we are doing classifier free guidance lowerCamelCase_ = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowerCamelCase_ = self.scheduler.scale_model_input(lowercase , lowercase ) lowerCamelCase_ = self.prior( lowercase , timestep=lowercase , proj_embedding=lowercase , ).predicted_image_embedding # remove the variance lowerCamelCase_ , lowerCamelCase_ = noise_pred.split( scaled_model_input.shape[2] , dim=2 ) # batch_size, num_embeddings, embedding_dim if do_classifier_free_guidance is not None: lowerCamelCase_ , lowerCamelCase_ = noise_pred.chunk(2 ) lowerCamelCase_ = noise_pred_uncond + guidance_scale * (noise_pred - noise_pred_uncond) lowerCamelCase_ = self.scheduler.step( lowercase , timestep=lowercase , sample=lowercase , ).prev_sample if output_type == "latent": return ShapEPipelineOutput(images=lowercase ) lowerCamelCase_ = [] for i, latent in enumerate(lowercase ): print() lowerCamelCase_ = self.renderer.decode( latent[None, :] , lowercase , size=lowercase , ray_batch_size=4096 , n_coarse_samples=64 , n_fine_samples=128 , ) images.append(lowercase ) lowerCamelCase_ = torch.stack(lowercase ) if output_type not in ["np", "pil"]: raise ValueError(f'Only the output types `pil` and `np` are supported not output_type={output_type}' ) lowerCamelCase_ = images.cpu().numpy() if output_type == "pil": lowerCamelCase_ = [self.numpy_to_pil(lowercase ) for image in images] # Offload last model to CPU if hasattr(self , "final_offload_hook" ) and self.final_offload_hook is not None: self.final_offload_hook.offload() if not return_dict: return (images,) return ShapEPipelineOutput(images=lowercase )
19
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_mbart import MBartTokenizer else: __A =None __A =logging.get_logger(__name__) __A ={'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} __A ={ '''vocab_file''': { '''facebook/mbart-large-en-ro''': ( '''https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model''' ), '''facebook/mbart-large-cc25''': ( '''https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/mbart-large-en-ro''': '''https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/tokenizer.json''', '''facebook/mbart-large-cc25''': '''https://huggingface.co/facebook/mbart-large-cc25/resolve/main/tokenizer.json''', }, } __A ={ '''facebook/mbart-large-en-ro''': 1_0_2_4, '''facebook/mbart-large-cc25''': 1_0_2_4, } # fmt: off __A =['''ar_AR''', '''cs_CZ''', '''de_DE''', '''en_XX''', '''es_XX''', '''et_EE''', '''fi_FI''', '''fr_XX''', '''gu_IN''', '''hi_IN''', '''it_IT''', '''ja_XX''', '''kk_KZ''', '''ko_KR''', '''lt_LT''', '''lv_LV''', '''my_MM''', '''ne_NP''', '''nl_XX''', '''ro_RO''', '''ru_RU''', '''si_LK''', '''tr_TR''', '''vi_VN''', '''zh_CN'''] class _SCREAMING_SNAKE_CASE ( snake_case_ ): lowerCAmelCase__ = VOCAB_FILES_NAMES lowerCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase__ = ['input_ids', 'attention_mask'] lowerCAmelCase__ = MBartTokenizer lowerCAmelCase__ = [] lowerCAmelCase__ = [] def __init__( self , lowercase=None , lowercase=None , lowercase="<s>" , lowercase="</s>" , lowercase="</s>" , lowercase="<s>" , lowercase="<unk>" , lowercase="<pad>" , lowercase="<mask>" , lowercase=None , lowercase=None , lowercase=None , **lowercase , ) -> Dict: # Mask token behave like a normal word, i.e. include the space before it lowerCamelCase_ = AddedToken(lowercase , lstrip=lowercase , rstrip=lowercase ) if isinstance(lowercase , lowercase ) else mask_token super().__init__( vocab_file=lowercase , tokenizer_file=lowercase , bos_token=lowercase , eos_token=lowercase , sep_token=lowercase , cls_token=lowercase , unk_token=lowercase , pad_token=lowercase , mask_token=lowercase , src_lang=lowercase , tgt_lang=lowercase , additional_special_tokens=lowercase , **lowercase , ) lowerCamelCase_ = vocab_file lowerCamelCase_ = False if not self.vocab_file else True lowerCamelCase_ = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({"additional_special_tokens": _additional_special_tokens} ) lowerCamelCase_ = { lang_code: self.convert_tokens_to_ids(lowercase ) for lang_code in FAIRSEQ_LANGUAGE_CODES } lowerCamelCase_ = src_lang if src_lang is not None else "en_XX" lowerCamelCase_ = self.convert_tokens_to_ids(self._src_lang ) lowerCamelCase_ = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def SCREAMING_SNAKE_CASE_( self ) -> str: return self._src_lang @src_lang.setter def SCREAMING_SNAKE_CASE_( self , lowercase ) -> None: lowerCamelCase_ = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase = None ) -> List[int]: lowerCamelCase_ = [self.sep_token_id] lowerCamelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , **lowercase ) -> List[Any]: if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model" ) lowerCamelCase_ = src_lang lowerCamelCase_ = self(lowercase , add_special_tokens=lowercase , return_tensors=lowercase , **lowercase ) lowerCamelCase_ = self.convert_tokens_to_ids(lowercase ) lowerCamelCase_ = tgt_lang_id return inputs def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase = "en_XX" , lowercase = None , lowercase = "ro_RO" , **lowercase , ) -> BatchEncoding: lowerCamelCase_ = src_lang lowerCamelCase_ = tgt_lang return super().prepare_seqaseq_batch(lowercase , lowercase , **lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> Dict: return self.set_src_lang_special_tokens(self.src_lang ) def SCREAMING_SNAKE_CASE_( self ) -> Optional[Any]: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def SCREAMING_SNAKE_CASE_( self , lowercase ) -> None: lowerCamelCase_ = self.convert_tokens_to_ids(lowercase ) lowerCamelCase_ = [] lowerCamelCase_ = [self.eos_token_id, self.cur_lang_code] lowerCamelCase_ = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCamelCase_ = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCamelCase_ = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def SCREAMING_SNAKE_CASE_( self , lowercase ) -> None: lowerCamelCase_ = self.convert_tokens_to_ids(lowercase ) lowerCamelCase_ = [] lowerCamelCase_ = [self.eos_token_id, self.cur_lang_code] lowerCamelCase_ = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCamelCase_ = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCamelCase_ = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowercase ): logger.error(f'Vocabulary path ({save_directory}) should be a directory.' ) return lowerCamelCase_ = os.path.join( lowercase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase ): copyfile(self.vocab_file , lowercase ) return (out_vocab_file,)
19
1
"""simple docstring""" from __future__ import annotations from collections import namedtuple def snake_case__ ( __lowerCamelCase : float , __lowerCamelCase : float , __lowerCamelCase : float ): """simple docstring""" lowerCamelCase__ : int =namedtuple('''result''' , '''name value''' ) if (voltage, current, power).count(0 ) != 1: raise ValueError('''Only one argument must be 0''' ) elif power < 0: raise ValueError( '''Power cannot be negative in any electrical/electronics system''' ) elif voltage == 0: return result('''voltage''' , power / current ) elif current == 0: return result('''current''' , power / voltage ) elif power == 0: return result('''power''' , float(round(abs(voltage * current ) , 2 ) ) ) else: raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
272
"""simple docstring""" from __future__ import annotations def snake_case__ ( __lowerCamelCase : str , __lowerCamelCase : list[str] | None = None ): """simple docstring""" lowerCamelCase__ : List[Any] =word_bank or [] # create a table lowerCamelCase__ : int =len(__lowerCamelCase ) + 1 lowerCamelCase__ : list[list[list[str]]] =[] for _ in range(__lowerCamelCase ): table.append([] ) # seed value lowerCamelCase__ : Tuple =[[]] # because empty string has empty combination # iterate through the indices for i in range(__lowerCamelCase ): # condition if table[i] != []: for word in word_bank: # slice condition if target[i : i + len(__lowerCamelCase )] == word: lowerCamelCase__ : list[list[str]] =[ [word, *way] for way in table[i] ] # adds the word to every combination the current position holds # now,push that combination to the table[i+len(word)] table[i + len(__lowerCamelCase )] += new_combinations # combinations are in reverse order so reverse for better output for combination in table[len(__lowerCamelCase )]: combination.reverse() return table[len(__lowerCamelCase )] if __name__ == "__main__": print(all_construct("jwajalapa", ["jwa", "j", "w", "a", "la", "lapa"])) print(all_construct("rajamati", ["s", "raj", "amat", "raja", "ma", "i", "t"])) print( all_construct( "hexagonosaurus", ["h", "ex", "hex", "ag", "ago", "ru", "auru", "rus", "go", "no", "o", "s"], ) )
272
1
def lowerCAmelCase__( lowercase : list ) -> list: for i in range(len(lowercase ) - 1 , 0 , -1 ): __snake_case : List[str] = False for j in range(lowercase , 0 , -1 ): if unsorted[j] < unsorted[j - 1]: __snake_case , __snake_case : Union[str, Any] = unsorted[j - 1], unsorted[j] __snake_case : Dict = True for j in range(lowercase ): if unsorted[j] > unsorted[j + 1]: __snake_case , __snake_case : Dict = unsorted[j + 1], unsorted[j] __snake_case : Any = True if not swapped: break return unsorted if __name__ == "__main__": import doctest doctest.testmod() _UpperCamelCase = input('''Enter numbers separated by a comma:\n''').strip() _UpperCamelCase = [int(item) for item in user_input.split(''',''')] print(F'''{cocktail_shaker_sort(unsorted) = }''')
326
def lowerCAmelCase__( lowercase : List[Any] , lowercase : Optional[Any] , lowercase : Optional[int] , lowercase : str , lowercase : List[Any] , lowercase : List[str] ) -> int: if index == r: for j in range(lowercase ): print(data[j] , end=" " ) print(" " ) return # When no more elements are there to put in data[] if i >= n: return # current is included, put next at next location __snake_case : Union[str, Any] = arr[i] combination_util(lowercase , lowercase , lowercase , index + 1 , lowercase , i + 1 ) # current is excluded, replace it with # next (Note that i+1 is passed, but # index is not changed) combination_util(lowercase , lowercase , lowercase , lowercase , lowercase , i + 1 ) # The main function that prints all combinations # of size r in arr[] of size n. This function # mainly uses combinationUtil() def lowerCAmelCase__( lowercase : Any , lowercase : Tuple , lowercase : Union[str, Any] ) -> Optional[Any]: # A temporary array to store all combination one by one __snake_case : Tuple = [0] * r # Print all combination using temporary array 'data[]' combination_util(lowercase , lowercase , lowercase , 0 , lowercase , 0 ) if __name__ == "__main__": # Driver code to check the function above _UpperCamelCase = [10, 20, 30, 40, 50] print_combination(arr, len(arr), 3) # This code is contributed by Ambuj sahu
326
1
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase : Optional[Any] = logging.get_logger(__name__) lowerCAmelCase : str = { """facebook/s2t-wav2vec2-large-en-de""": ( """https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/config.json""" ), # See all Speech2Text models at https://huggingface.co/models?filter=speech2text2 } class __lowercase ( UpperCAmelCase_ ): """simple docstring""" _UpperCAmelCase : Optional[Any] = '''speech_to_text_2''' _UpperCAmelCase : Optional[int] = ['''past_key_values'''] _UpperCAmelCase : Dict = {'''num_attention_heads''': '''decoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : Tuple , lowerCAmelCase__ : Dict=1_0000 , lowerCAmelCase__ : Dict=6 , lowerCAmelCase__ : Dict=2048 , lowerCAmelCase__ : List[str]=4 , lowerCAmelCase__ : Any=0.0 , lowerCAmelCase__ : int=True , lowerCAmelCase__ : Optional[int]="relu" , lowerCAmelCase__ : str=256 , lowerCAmelCase__ : List[str]=0.1 , lowerCAmelCase__ : Dict=0.0 , lowerCAmelCase__ : str=0.0 , lowerCAmelCase__ : str=0.02 , lowerCAmelCase__ : Dict=2 , lowerCAmelCase__ : Optional[int]=True , lowerCAmelCase__ : Optional[int]=1 , lowerCAmelCase__ : Optional[int]=0 , lowerCAmelCase__ : Optional[int]=2 , lowerCAmelCase__ : int=1024 , **lowerCAmelCase__ : Optional[Any] , ): SCREAMING_SNAKE_CASE_: Optional[Any] = vocab_size SCREAMING_SNAKE_CASE_: Optional[Any] = d_model SCREAMING_SNAKE_CASE_: str = decoder_ffn_dim SCREAMING_SNAKE_CASE_: str = decoder_layers SCREAMING_SNAKE_CASE_: int = decoder_attention_heads SCREAMING_SNAKE_CASE_: Tuple = dropout SCREAMING_SNAKE_CASE_: Union[str, Any] = attention_dropout SCREAMING_SNAKE_CASE_: Optional[int] = activation_dropout SCREAMING_SNAKE_CASE_: List[str] = activation_function SCREAMING_SNAKE_CASE_: Union[str, Any] = init_std SCREAMING_SNAKE_CASE_: Tuple = decoder_layerdrop SCREAMING_SNAKE_CASE_: List[Any] = use_cache SCREAMING_SNAKE_CASE_: List[str] = decoder_layers SCREAMING_SNAKE_CASE_: Optional[int] = scale_embedding # scale factor will be sqrt(d_model) if True SCREAMING_SNAKE_CASE_: List[Any] = max_target_positions super().__init__( pad_token_id=lowerCAmelCase__ , bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , decoder_start_token_id=lowerCAmelCase__ , **lowerCAmelCase__ , )
127
import re from typing import Callable, List, Optional, Union import tensorflow as tf try: from tensorflow.keras.optimizers.legacy import Adam except ImportError: from tensorflow.keras.optimizers import Adam class __lowercase ( tf.keras.optimizers.schedules.LearningRateSchedule ): """simple docstring""" def __init__( self : List[Any] , lowerCAmelCase__ : float , lowerCAmelCase__ : Callable , lowerCAmelCase__ : int , lowerCAmelCase__ : float = 1.0 , lowerCAmelCase__ : str = None , ): super().__init__() SCREAMING_SNAKE_CASE_: str = initial_learning_rate SCREAMING_SNAKE_CASE_: Dict = warmup_steps SCREAMING_SNAKE_CASE_: Any = power SCREAMING_SNAKE_CASE_: int = decay_schedule_fn SCREAMING_SNAKE_CASE_: Union[str, Any] = name def __call__( self : Optional[Any] , lowerCAmelCase__ : Any): with tf.name_scope(self.name or "WarmUp") as name: # Implements polynomial warmup. i.e., if global_step < warmup_steps, the # learning rate will be `global_step/num_warmup_steps * init_lr`. SCREAMING_SNAKE_CASE_: Any = tf.cast(lowerCAmelCase__ , tf.floataa) SCREAMING_SNAKE_CASE_: Optional[Any] = tf.cast(self.warmup_steps , tf.floataa) SCREAMING_SNAKE_CASE_: Optional[int] = global_step_float / warmup_steps_float SCREAMING_SNAKE_CASE_: Union[str, Any] = self.initial_learning_rate * tf.math.pow(lowerCAmelCase__ , self.power) return tf.cond( global_step_float < warmup_steps_float , lambda: warmup_learning_rate , lambda: self.decay_schedule_fn(step - self.warmup_steps) , name=lowerCAmelCase__ , ) def _SCREAMING_SNAKE_CASE ( self : Tuple): return { "initial_learning_rate": self.initial_learning_rate, "decay_schedule_fn": self.decay_schedule_fn, "warmup_steps": self.warmup_steps, "power": self.power, "name": self.name, } def A_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = 0.0 , _UpperCAmelCase = 0.9 , _UpperCAmelCase = 0.9_9_9 , _UpperCAmelCase = 1e-8 , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = 0.0 , _UpperCAmelCase = 1.0 , _UpperCAmelCase = None , ): SCREAMING_SNAKE_CASE_: Optional[int] = tf.keras.optimizers.schedules.PolynomialDecay( initial_learning_rate=_UpperCAmelCase , decay_steps=num_train_steps - num_warmup_steps , end_learning_rate=init_lr * min_lr_ratio , power=_UpperCAmelCase , ) if num_warmup_steps: SCREAMING_SNAKE_CASE_: Tuple = WarmUp( initial_learning_rate=_UpperCAmelCase , decay_schedule_fn=_UpperCAmelCase , warmup_steps=_UpperCAmelCase , ) if weight_decay_rate > 0.0: SCREAMING_SNAKE_CASE_: List[str] = AdamWeightDecay( learning_rate=_UpperCAmelCase , weight_decay_rate=_UpperCAmelCase , beta_a=_UpperCAmelCase , beta_a=_UpperCAmelCase , epsilon=_UpperCAmelCase , clipnorm=_UpperCAmelCase , global_clipnorm=_UpperCAmelCase , exclude_from_weight_decay=["LayerNorm", "layer_norm", "bias"] , include_in_weight_decay=_UpperCAmelCase , ) else: SCREAMING_SNAKE_CASE_: int = tf.keras.optimizers.Adam( learning_rate=_UpperCAmelCase , beta_a=_UpperCAmelCase , beta_a=_UpperCAmelCase , epsilon=_UpperCAmelCase , clipnorm=_UpperCAmelCase , global_clipnorm=_UpperCAmelCase , ) # We return the optimizer and the LR scheduler in order to better track the # evolution of the LR independently of the optimizer. return optimizer, lr_schedule class __lowercase ( UpperCAmelCase_ ): """simple docstring""" def __init__( self : List[Any] , lowerCAmelCase__ : Union[float, tf.keras.optimizers.schedules.LearningRateSchedule] = 0.001 , lowerCAmelCase__ : float = 0.9 , lowerCAmelCase__ : float = 0.999 , lowerCAmelCase__ : float = 1E-7 , lowerCAmelCase__ : bool = False , lowerCAmelCase__ : float = 0.0 , lowerCAmelCase__ : Optional[List[str]] = None , lowerCAmelCase__ : Optional[List[str]] = None , lowerCAmelCase__ : str = "AdamWeightDecay" , **lowerCAmelCase__ : int , ): super().__init__(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Optional[Any] = weight_decay_rate SCREAMING_SNAKE_CASE_: List[Any] = include_in_weight_decay SCREAMING_SNAKE_CASE_: List[Any] = exclude_from_weight_decay @classmethod def _SCREAMING_SNAKE_CASE ( cls : Dict , lowerCAmelCase__ : Dict): SCREAMING_SNAKE_CASE_: List[str] = {"WarmUp": WarmUp} return super(lowerCAmelCase__ , cls).from_config(lowerCAmelCase__ , custom_objects=lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : List[str] , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Optional[int]): super(lowerCAmelCase__ , self)._prepare_local(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Optional[int] = tf.constant( self.weight_decay_rate , name="adam_weight_decay_rate") def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , lowerCAmelCase__ : str , lowerCAmelCase__ : Any , lowerCAmelCase__ : Tuple): SCREAMING_SNAKE_CASE_: str = self._do_use_weight_decay(var.name) if do_decay: return var.assign_sub( learning_rate * var * apply_state[(var.device, var.dtype.base_dtype)]["weight_decay_rate"] , use_locking=self._use_locking , ) return tf.no_op() def _SCREAMING_SNAKE_CASE ( self : Dict , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : Optional[Any]=None , **lowerCAmelCase__ : List[str]): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Any = list(zip(*lowerCAmelCase__)) return super(lowerCAmelCase__ , self).apply_gradients(zip(lowerCAmelCase__ , lowerCAmelCase__) , name=lowerCAmelCase__ , **lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : Optional[int] , lowerCAmelCase__ : Any , lowerCAmelCase__ : Tuple , lowerCAmelCase__ : Tuple): if apply_state is None: return self._decayed_lr_t[var_dtype], {} SCREAMING_SNAKE_CASE_: Dict = apply_state or {} SCREAMING_SNAKE_CASE_: List[str] = apply_state.get((var_device, var_dtype)) if coefficients is None: SCREAMING_SNAKE_CASE_: Optional[int] = self._fallback_apply_state(lowerCAmelCase__ , lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Any = coefficients return coefficients["lr_t"], {"apply_state": apply_state} def _SCREAMING_SNAKE_CASE ( self : str , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Dict , lowerCAmelCase__ : Tuple=None): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Optional[Any] = self._get_lr(var.device , var.dtype.base_dtype , lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Optional[int] = self._decay_weights_op(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__) with tf.control_dependencies([decay]): return super(lowerCAmelCase__ , self)._resource_apply_dense(lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Dict , lowerCAmelCase__ : Dict , lowerCAmelCase__ : Dict=None): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: List[str] = self._get_lr(var.device , var.dtype.base_dtype , lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Any = self._decay_weights_op(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__) with tf.control_dependencies([decay]): return super(lowerCAmelCase__ , self)._resource_apply_sparse(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): SCREAMING_SNAKE_CASE_: List[str] = super().get_config() config.update({"weight_decay_rate": self.weight_decay_rate}) return config def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] , lowerCAmelCase__ : Tuple): if self.weight_decay_rate == 0: return False if self._include_in_weight_decay: for r in self._include_in_weight_decay: if re.search(lowerCAmelCase__ , lowerCAmelCase__) is not None: return True if self._exclude_from_weight_decay: for r in self._exclude_from_weight_decay: if re.search(lowerCAmelCase__ , lowerCAmelCase__) is not None: return False return True class __lowercase ( UpperCAmelCase_ ): """simple docstring""" def __init__( self : Optional[Any]): SCREAMING_SNAKE_CASE_: Any = [] SCREAMING_SNAKE_CASE_: Any = None @property def _SCREAMING_SNAKE_CASE ( self : int): if self._accum_steps is None: SCREAMING_SNAKE_CASE_: Tuple = tf.Variable( tf.constant(0 , dtype=tf.intaa) , trainable=lowerCAmelCase__ , synchronization=tf.VariableSynchronization.ON_READ , aggregation=tf.VariableAggregation.ONLY_FIRST_REPLICA , ) return self._accum_steps.value() @property def _SCREAMING_SNAKE_CASE ( self : Tuple): if not self._gradients: raise ValueError("The accumulator should be called first to initialize the gradients") return [gradient.value() if gradient is not None else gradient for gradient in self._gradients] def __call__( self : str , lowerCAmelCase__ : Tuple): if not self._gradients: SCREAMING_SNAKE_CASE_: Optional[Any] = self.step # Create the step variable. self._gradients.extend( [ tf.Variable( tf.zeros_like(lowerCAmelCase__) , trainable=lowerCAmelCase__ , synchronization=tf.VariableSynchronization.ON_READ , aggregation=tf.VariableAggregation.ONLY_FIRST_REPLICA , ) if gradient is not None else gradient for gradient in gradients ]) if len(lowerCAmelCase__) != len(self._gradients): raise ValueError(F"Expected {len(self._gradients)} gradients, but got {len(lowerCAmelCase__)}") for accum_gradient, gradient in zip(self._gradients , lowerCAmelCase__): if accum_gradient is not None and gradient is not None: accum_gradient.assign_add(lowerCAmelCase__) self._accum_steps.assign_add(1) def _SCREAMING_SNAKE_CASE ( self : int): if not self._gradients: return self._accum_steps.assign(0) for gradient in self._gradients: if gradient is not None: gradient.assign(tf.zeros_like(lowerCAmelCase__))
127
1
# flake8: noqa # Lint as: python3 from typing import Dict, List, Optional, Type from .. import config from ..utils import logging from .formatting import ( ArrowFormatter, CustomFormatter, Formatter, PandasFormatter, PythonFormatter, TensorFormatter, format_table, query_table, ) from .np_formatter import NumpyFormatter lowercase : List[str] = logging.get_logger(__name__) lowercase : Dict[Optional[str], Type[Formatter]] = {} lowercase : Dict[Optional[str], str] = {} lowercase : Dict[Optional[str], Exception] = {} def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None , ) -> Optional[int]: lowercase : Optional[int] = aliases if aliases is not None else [] if format_type in _FORMAT_TYPES: logger.warning( f"Overwriting format type \'{format_type}\' ({_FORMAT_TYPES[format_type].__name__} -> {formatter_cls.__name__})" ) lowercase : int = formatter_cls for alias in set(aliases + [format_type] ): if alias in _FORMAT_TYPES_ALIASES: logger.warning( f"Overwriting format type alias \'{alias}\' ({_FORMAT_TYPES_ALIASES[alias]} -> {format_type})" ) lowercase : Union[str, Any] = format_type def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None ) -> List[str]: lowercase : Any = aliases if aliases is not None else [] for alias in set(aliases + [format_type] ): lowercase : int = unavailable_error # Here we define all the available formatting functions that can be used by `Dataset.set_format` _register_formatter(PythonFormatter, None, aliases=["""python"""]) _register_formatter(ArrowFormatter, """arrow""", aliases=["""pa""", """pyarrow"""]) _register_formatter(NumpyFormatter, """numpy""", aliases=["""np"""]) _register_formatter(PandasFormatter, """pandas""", aliases=["""pd"""]) _register_formatter(CustomFormatter, """custom""") if config.TORCH_AVAILABLE: from .torch_formatter import TorchFormatter _register_formatter(TorchFormatter, """torch""", aliases=["""pt""", """pytorch"""]) else: lowercase : str = ValueError("""PyTorch needs to be installed to be able to return PyTorch tensors.""") _register_unavailable_formatter(_torch_error, """torch""", aliases=["""pt""", """pytorch"""]) if config.TF_AVAILABLE: from .tf_formatter import TFFormatter _register_formatter(TFFormatter, """tensorflow""", aliases=["""tf"""]) else: lowercase : List[str] = ValueError("""Tensorflow needs to be installed to be able to return Tensorflow tensors.""") _register_unavailable_formatter(_tf_error, """tensorflow""", aliases=["""tf"""]) if config.JAX_AVAILABLE: from .jax_formatter import JaxFormatter _register_formatter(JaxFormatter, """jax""", aliases=[]) else: lowercase : List[str] = ValueError("""JAX needs to be installed to be able to return JAX arrays.""") _register_unavailable_formatter(_jax_error, """jax""", aliases=[]) def _snake_case( SCREAMING_SNAKE_CASE__ ) -> Optional[str]: if format_type in _FORMAT_TYPES_ALIASES: return _FORMAT_TYPES_ALIASES[format_type] else: return format_type def _snake_case( SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) -> Formatter: lowercase : List[str] = get_format_type_from_alias(__lowerCamelCase ) if format_type in _FORMAT_TYPES: return _FORMAT_TYPES[format_type](**__lowerCamelCase ) if format_type in _FORMAT_TYPES_ALIASES_UNAVAILABLE: raise _FORMAT_TYPES_ALIASES_UNAVAILABLE[format_type] else: raise ValueError( f"Return type should be None or selected in {list(type for type in _FORMAT_TYPES.keys() if type != None )}, but got \'{format_type}\'" )
20
import json import os import unittest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __lowerCAmelCase ( __magic_name__ , unittest.TestCase ): UpperCamelCase__ = CLIPTokenizer UpperCamelCase__ = CLIPTokenizerFast UpperCamelCase__ = True UpperCamelCase__ = {} UpperCamelCase__ = False def lowerCamelCase__ ( self :str ): '''simple docstring''' super().setUp() # fmt: off a = ["""l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """lo""", """l</w>""", """w</w>""", """r</w>""", """t</w>""", """low</w>""", """er</w>""", """lowest</w>""", """newer</w>""", """wider""", """<unk>""", """<|startoftext|>""", """<|endoftext|>"""] # fmt: on a = dict(zip(__magic_name__ , range(len(__magic_name__ ) ) ) ) a = ["""#version: 0.2""", """l o""", """lo w</w>""", """e r</w>"""] a = {"""unk_token""": """<unk>"""} a = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) a = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__magic_name__ ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(__magic_name__ ) ) def lowerCamelCase__ ( self :Optional[Any] , **__magic_name__ :Union[str, Any] ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return CLIPTokenizer.from_pretrained(self.tmpdirname , **__magic_name__ ) def lowerCamelCase__ ( self :int , **__magic_name__ :Optional[Any] ): '''simple docstring''' kwargs.update(self.special_tokens_map ) return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **__magic_name__ ) def lowerCamelCase__ ( self :int , __magic_name__ :List[str] ): '''simple docstring''' a = """lower newer""" a = """lower newer""" return input_text, output_text def lowerCamelCase__ ( self :int ): '''simple docstring''' a = CLIPTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) a = """lower newer""" a = ["""lo""", """w""", """er</w>""", """n""", """e""", """w""", """er</w>"""] a = tokenizer.tokenize(__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) a = tokens + [tokenizer.unk_token] a = [10, 2, 16, 9, 3, 2, 16, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(__magic_name__ ) , __magic_name__ ) @require_ftfy def lowerCamelCase__ ( self :List[Any] ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): a = self.tokenizer_class.from_pretrained(__magic_name__ , **__magic_name__ ) a = self.rust_tokenizer_class.from_pretrained(__magic_name__ , **__magic_name__ ) a = """A\n'll 11p223RF☆ho!!to?'d'd''d of a cat to-$''d.""" a = tokenizer_s.tokenize(__magic_name__ ) a = tokenizer_r.tokenize(__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) # Test that the tokenization is identical on an example containing a character (Latin Small Letter A # with Tilde) encoded in 2 different ways a = """xa\u0303y""" + """ """ + """x\xe3y""" a = tokenizer_s.tokenize(__magic_name__ ) a = tokenizer_r.tokenize(__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) # Test that the tokenization is identical on unicode of space type a = [ """\u0009""", # (horizontal tab, '\t') """\u000B""", # (vertical tab) """\u000C""", # (form feed) """\u0020""", # (space, ' ') """\u200E""", # (left-to-right mark):w """\u200F""", # (right-to-left mark) ] for unicode_seq in spaces_unicodes: a = tokenizer_s.tokenize(__magic_name__ ) a = tokenizer_r.tokenize(__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) # Test that the tokenization is identical on unicode of line break type a = [ """\u000A""", # (line feed, '\n') """\r\n""", # (carriage return and line feed, '\r\n') """\u000D""", # (carriage return, '\r') """\r""", # (carriage return, '\r') """\u000D""", # (carriage return, '\r') """\u2028""", # (line separator) """\u2029""", # (paragraph separator) # "\u0085", # (next line) ] # The tokenization is not identical for the character "\u0085" (next line). The slow version using ftfy transforms # it into the Horizontal Ellipsis character "…" ("\u2026") while the fast version transforms it into a # space (and thus into an empty list). for unicode_seq in line_break_unicodes: a = tokenizer_s.tokenize(__magic_name__ ) a = tokenizer_r.tokenize(__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) def lowerCamelCase__ ( self :int ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): a = """hello""" # `hello` is a token in the vocabulary of `pretrained_name` a = F'{text_of_1_token} {text_of_1_token}' a = self.rust_tokenizer_class.from_pretrained( __magic_name__ , use_fast=__magic_name__ , ) a = tokenizer_r(__magic_name__ , return_offsets_mapping=__magic_name__ , add_special_tokens=__magic_name__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(__magic_name__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(__magic_name__ ) + 1, len(__magic_name__ ) + 1 + len(__magic_name__ )) , ) a = F' {text}' a = self.rust_tokenizer_class.from_pretrained( __magic_name__ , use_fast=__magic_name__ , ) a = tokenizer_r(__magic_name__ , return_offsets_mapping=__magic_name__ , add_special_tokens=__magic_name__ ) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(__magic_name__ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(__magic_name__ ) + 1, 1 + len(__magic_name__ ) + 1 + len(__magic_name__ )) , ) def lowerCamelCase__ ( self :List[str] ): '''simple docstring''' with self.assertRaises(__magic_name__ ) as context: self.rust_tokenizer_class.from_pretrained("""robot-test/old-clip-tokenizer""" ) self.assertTrue( context.exception.args[0].startswith( """The `backend_tokenizer` provided does not match the expected format.""" ) ) @require_ftfy def lowerCamelCase__ ( self :Tuple ): '''simple docstring''' super().test_tokenization_python_rust_equals() def lowerCamelCase__ ( self :Optional[int] ): '''simple docstring''' pass
228
0
from ..utils import DummyObject, requires_backends class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: int = ["sentencepiece"] def __init__( self: Dict , *__lowerCamelCase: Optional[int] , **__lowerCamelCase: Union[str, Any] ) -> List[str]: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: str = ["sentencepiece"] def __init__( self: int , *__lowerCamelCase: Optional[int] , **__lowerCamelCase: Tuple ) -> List[str]: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: int = ["sentencepiece"] def __init__( self: Union[str, Any] , *__lowerCamelCase: List[str] , **__lowerCamelCase: Any ) -> Any: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: Any = ["sentencepiece"] def __init__( self: Union[str, Any] , *__lowerCamelCase: Any , **__lowerCamelCase: Optional[Any] ) -> int: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: Union[str, Any] = ["sentencepiece"] def __init__( self: Dict , *__lowerCamelCase: List[str] , **__lowerCamelCase: Optional[Any] ) -> Union[str, Any]: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: Optional[Any] = ["sentencepiece"] def __init__( self: Tuple , *__lowerCamelCase: Optional[Any] , **__lowerCamelCase: Tuple ) -> Optional[int]: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: List[str] = ["sentencepiece"] def __init__( self: Any , *__lowerCamelCase: Optional[Any] , **__lowerCamelCase: Union[str, Any] ) -> Dict: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: Tuple = ["sentencepiece"] def __init__( self: int , *__lowerCamelCase: str , **__lowerCamelCase: Any ) -> Any: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: str = ["sentencepiece"] def __init__( self: Tuple , *__lowerCamelCase: str , **__lowerCamelCase: Tuple ) -> List[str]: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: Dict = ["sentencepiece"] def __init__( self: List[str] , *__lowerCamelCase: Any , **__lowerCamelCase: int ) -> List[Any]: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: Union[str, Any] = ["sentencepiece"] def __init__( self: Optional[Any] , *__lowerCamelCase: Union[str, Any] , **__lowerCamelCase: List[Any] ) -> str: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: str = ["sentencepiece"] def __init__( self: int , *__lowerCamelCase: List[Any] , **__lowerCamelCase: List[Any] ) -> List[Any]: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: List[str] = ["sentencepiece"] def __init__( self: Dict , *__lowerCamelCase: List[str] , **__lowerCamelCase: List[str] ) -> Optional[Any]: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: int = ["sentencepiece"] def __init__( self: Optional[Any] , *__lowerCamelCase: Optional[int] , **__lowerCamelCase: Tuple ) -> Optional[int]: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: Union[str, Any] = ["sentencepiece"] def __init__( self: Optional[Any] , *__lowerCamelCase: Optional[Any] , **__lowerCamelCase: Tuple ) -> Optional[Any]: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: Union[str, Any] = ["sentencepiece"] def __init__( self: str , *__lowerCamelCase: Dict , **__lowerCamelCase: List[Any] ) -> Union[str, Any]: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: Tuple = ["sentencepiece"] def __init__( self: Optional[Any] , *__lowerCamelCase: Union[str, Any] , **__lowerCamelCase: Tuple ) -> Tuple: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: Any = ["sentencepiece"] def __init__( self: List[str] , *__lowerCamelCase: int , **__lowerCamelCase: Dict ) -> str: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: List[str] = ["sentencepiece"] def __init__( self: Tuple , *__lowerCamelCase: List[Any] , **__lowerCamelCase: Dict ) -> int: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: List[str] = ["sentencepiece"] def __init__( self: Any , *__lowerCamelCase: List[Any] , **__lowerCamelCase: int ) -> Tuple: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: Union[str, Any] = ["sentencepiece"] def __init__( self: Optional[int] , *__lowerCamelCase: Union[str, Any] , **__lowerCamelCase: List[str] ) -> Union[str, Any]: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: Union[str, Any] = ["sentencepiece"] def __init__( self: Union[str, Any] , *__lowerCamelCase: Dict , **__lowerCamelCase: Optional[int] ) -> List[Any]: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: Optional[int] = ["sentencepiece"] def __init__( self: str , *__lowerCamelCase: str , **__lowerCamelCase: Optional[int] ) -> Union[str, Any]: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: str = ["sentencepiece"] def __init__( self: Any , *__lowerCamelCase: Union[str, Any] , **__lowerCamelCase: Optional[Any] ) -> List[Any]: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: Tuple = ["sentencepiece"] def __init__( self: int , *__lowerCamelCase: List[str] , **__lowerCamelCase: List[str] ) -> Any: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: Tuple = ["sentencepiece"] def __init__( self: Tuple , *__lowerCamelCase: Optional[Any] , **__lowerCamelCase: Tuple ) -> List[str]: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: int = ["sentencepiece"] def __init__( self: List[str] , *__lowerCamelCase: Tuple , **__lowerCamelCase: str ) -> str: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: List[str] = ["sentencepiece"] def __init__( self: Union[str, Any] , *__lowerCamelCase: int , **__lowerCamelCase: Dict ) -> str: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: Optional[int] = ["sentencepiece"] def __init__( self: List[Any] , *__lowerCamelCase: Tuple , **__lowerCamelCase: Optional[Any] ) -> List[Any]: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: int = ["sentencepiece"] def __init__( self: List[str] , *__lowerCamelCase: Any , **__lowerCamelCase: Optional[Any] ) -> Optional[Any]: requires_backends(self , ["sentencepiece"] ) class _snake_case ( metaclass=_lowercase ): lowerCamelCase__: List[Any] = ["sentencepiece"] def __init__( self: int , *__lowerCamelCase: Union[str, Any] , **__lowerCamelCase: Optional[int] ) -> Any: requires_backends(self , ["sentencepiece"] )
351
def _UpperCamelCase ( snake_case__ ) -> int: __UpperCAmelCase : int = [0] * len(snake_case__ ) __UpperCAmelCase : Union[str, Any] = [] __UpperCAmelCase : str = [1] * len(snake_case__ ) for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(snake_case__ ) ): if indegree[i] == 0: queue.append(snake_case__ ) while queue: __UpperCAmelCase : List[str] = queue.pop(0 ) for x in graph[vertex]: indegree[x] -= 1 if long_dist[vertex] + 1 > long_dist[x]: __UpperCAmelCase : str = long_dist[vertex] + 1 if indegree[x] == 0: queue.append(snake_case__ ) print(max(snake_case__ ) ) # Adjacency list of Graph _snake_case = {0: [2, 3, 4], 1: [2, 7], 2: [5], 3: [5, 7], 4: [7], 5: [6], 6: [7], 7: []} longest_distance(graph)
342
0
import os import unittest from transformers import LxmertTokenizer, LxmertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowercase__ ( _UpperCAmelCase , unittest.TestCase ): A__ : Dict =LxmertTokenizer A__ : List[Any] =LxmertTokenizerFast A__ : Any =True A__ : List[Any] =True def A_ ( self : Optional[Any] ): super().setUp() SCREAMING_SNAKE_CASE__ = [ '[UNK]', '[CLS]', '[SEP]', 'want', '##want', '##ed', 'wa', 'un', 'runn', '##ing', ',', 'low', 'lowest', ] SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) def A_ ( self : int , UpperCAmelCase_ : int ): SCREAMING_SNAKE_CASE__ = 'UNwant\u00E9d,running' SCREAMING_SNAKE_CASE__ = 'unwanted, running' return input_text, output_text def A_ ( self : Optional[Any] ): SCREAMING_SNAKE_CASE__ = self.tokenizer_class(self.vocab_file ) SCREAMING_SNAKE_CASE__ = tokenizer.tokenize('UNwant\u00E9d,running' ) self.assertListEqual(UpperCAmelCase_ , ['un', '##want', '##ed', ',', 'runn', '##ing'] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [7, 4, 5, 10, 8, 9] ) def A_ ( self : List[str] ): if not self.test_rust_tokenizer: return SCREAMING_SNAKE_CASE__ = self.get_tokenizer() SCREAMING_SNAKE_CASE__ = self.get_rust_tokenizer() SCREAMING_SNAKE_CASE__ = 'I was born in 92000, and this is falsé.' SCREAMING_SNAKE_CASE__ = tokenizer.tokenize(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = rust_tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = rust_tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = self.get_rust_tokenizer() SCREAMING_SNAKE_CASE__ = tokenizer.encode(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = rust_tokenizer.encode(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ )
176
import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": __snake_case = argparse.ArgumentParser() parser.add_argument( """--checkpoint_path""", default=None, type=str, required=True, help="""Path to the checkpoint to convert.""" ) parser.add_argument( """--original_config_file""", type=str, required=True, help="""The YAML config file corresponding to the original architecture.""", ) parser.add_argument( """--num_in_channels""", default=None, type=int, help="""The number of input channels. If `None` number of input channels will be automatically inferred.""", ) parser.add_argument( """--image_size""", default=5_12, type=int, help=( """The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2""" """ Base. Use 768 for Stable Diffusion v2.""" ), ) parser.add_argument( """--extract_ema""", action="""store_true""", help=( """Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights""" """ or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield""" """ higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning.""" ), ) parser.add_argument( """--upcast_attention""", action="""store_true""", help=( """Whether the attention computation should always be upcasted. This is necessary when running stable""" """ diffusion 2.1.""" ), ) parser.add_argument( """--from_safetensors""", action="""store_true""", help="""If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.""", ) parser.add_argument( """--to_safetensors""", action="""store_true""", help="""Whether to store pipeline in safetensors format or not.""", ) parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the output model.""") parser.add_argument("""--device""", type=str, help="""Device to use (e.g. cpu, cuda:0, cuda:1, etc.)""") def _lowercase ( UpperCamelCase_ ) -> Dict: '''simple docstring''' if string == "True": return True elif string == "False": return False else: raise ValueError(F'could not parse string as bool {string}' ) parser.add_argument( """--use_linear_projection""", help="""Override for use linear projection""", required=False, type=parse_bool ) parser.add_argument("""--cross_attention_dim""", help="""Override for cross attention_dim""", required=False, type=int) __snake_case = parser.parse_args() __snake_case = download_controlnet_from_original_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, extract_ema=args.extract_ema, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, use_linear_projection=args.use_linear_projection, cross_attention_dim=args.cross_attention_dim, ) controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
176
1
"""simple docstring""" __A = [ [0, 16, 13, 0, 0, 0], [0, 0, 10, 12, 0, 0], [0, 4, 0, 0, 14, 0], [0, 0, 9, 0, 0, 20], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] def lowercase_ ( _lowerCamelCase: Any , _lowerCamelCase: Optional[int] , _lowerCamelCase: Any , _lowerCamelCase: Union[str, Any] ) -> Optional[int]: '''simple docstring''' __lowerCamelCase : Dict = [False] * len(_lowerCamelCase ) __lowerCamelCase : Any = [s] __lowerCamelCase : Dict = True while queue: __lowerCamelCase : List[Any] = queue.pop(0 ) for ind in range(len(graph[u] ) ): if visited[ind] is False and graph[u][ind] > 0: queue.append(_lowerCamelCase ) __lowerCamelCase : Optional[Any] = True __lowerCamelCase : List[str] = u return visited[t] def lowercase_ ( _lowerCamelCase: Any , _lowerCamelCase: Union[str, Any] , _lowerCamelCase: Optional[Any] ) -> Union[str, Any]: '''simple docstring''' __lowerCamelCase : Optional[Any] = [-1] * (len(_lowerCamelCase )) __lowerCamelCase : Optional[Any] = 0 __lowerCamelCase : Optional[int] = [] __lowerCamelCase : List[Any] = [i[:] for i in graph] # Record original cut, copy. while bfs(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): __lowerCamelCase : List[Any] = float("Inf" ) __lowerCamelCase : List[str] = sink while s != source: # Find the minimum value in select path __lowerCamelCase : int = min(_lowerCamelCase , graph[parent[s]][s] ) __lowerCamelCase : Tuple = parent[s] max_flow += path_flow __lowerCamelCase : Optional[int] = sink while v != source: __lowerCamelCase : str = parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow __lowerCamelCase : Tuple = parent[v] for i in range(len(_lowerCamelCase ) ): for j in range(len(graph[0] ) ): if graph[i][j] == 0 and temp[i][j] > 0: res.append((i, j) ) return res if __name__ == "__main__": print(mincut(test_graph, source=0, sink=5))
64
"""simple docstring""" def lowercase_ ( _lowerCamelCase: int = 4000000 ) -> int: '''simple docstring''' __lowerCamelCase : Tuple = [0, 1] __lowerCamelCase : Union[str, Any] = 0 while fib[i] <= n: fib.append(fib[i] + fib[i + 1] ) if fib[i + 2] > n: break i += 1 __lowerCamelCase : Tuple = 0 for j in range(len(_lowerCamelCase ) - 1 ): if fib[j] % 2 == 0: total += fib[j] return total if __name__ == "__main__": print(F"""{solution() = }""")
64
1
'''simple docstring''' def snake_case_ ( _lowerCAmelCase : list[int] , _lowerCAmelCase : str ) -> list[int]: UpperCAmelCase : Optional[Any] = int(_lowerCAmelCase ) # Initialize Result UpperCAmelCase : List[Any] = [] # Traverse through all denomination for denomination in reversed(_lowerCAmelCase ): # Find denominations while int(_lowerCAmelCase ) >= int(_lowerCAmelCase ): total_value -= int(_lowerCAmelCase ) answer.append(_lowerCAmelCase ) # Append the "answers" array return answer # Driver Code if __name__ == "__main__": UpperCamelCase__: int = [] UpperCamelCase__: Optional[Any] = "0" if ( input("Do you want to enter your denominations ? (yY/n): ").strip().lower() == "y" ): UpperCamelCase__: int = int(input("Enter the number of denominations you want to add: ").strip()) for i in range(0, n): denominations.append(int(input(F"Denomination {i}: ").strip())) UpperCamelCase__: int = input("Enter the change you want to make in Indian Currency: ").strip() else: # All denominations of Indian Currency if user does not enter UpperCamelCase__: Any = [1, 2, 5, 10, 20, 50, 100, 500, 2000] UpperCamelCase__: Any = input("Enter the change you want to make: ").strip() if int(value) == 0 or int(value) < 0: print("The total value cannot be zero or negative.") else: print(F"Following is minimal change for {value}: ") UpperCamelCase__: Union[str, Any] = find_minimum_change(denominations, value) # Print result for i in range(len(answer)): print(answer[i], end=" ")
23
from typing import Dict import numpy as np import torch from . import residue_constants as rc from .tensor_utils import tensor_tree_map, tree_map def a ( lowerCamelCase_ ): '''simple docstring''' lowercase__ = [] lowercase__ = [] lowercase__ = [] for rt in rc.restypes: lowercase__ = rc.restype_name_to_atomaa_names[rc.restype_atoa[rt]] restype_atomaa_to_atomaa_list.append([(rc.atom_order[name] if name else 0) for name in atom_names] ) lowercase__ = {name: i for i, name in enumerate(lowerCamelCase_ )} restype_atomaa_to_atomaa_list.append( [(atom_name_to_idxaa[name] if name in atom_name_to_idxaa else 0) for name in rc.atom_types] ) restype_atomaa_mask_list.append([(1.0 if name else 0.0) for name in atom_names] ) # Add dummy mapping for restype 'UNK' restype_atomaa_to_atomaa_list.append([0] * 14 ) restype_atomaa_to_atomaa_list.append([0] * 37 ) restype_atomaa_mask_list.append([0.0] * 14 ) lowercase__ = torch.tensor( lowerCamelCase_ , dtype=torch.intaa , device=protein['''aatype'''].device , ) lowercase__ = torch.tensor( lowerCamelCase_ , dtype=torch.intaa , device=protein['''aatype'''].device , ) lowercase__ = torch.tensor( lowerCamelCase_ , dtype=torch.floataa , device=protein['''aatype'''].device , ) lowercase__ = protein['''aatype'''].to(torch.long ) # create the mapping for (residx, atom14) --> atom37, i.e. an array # with shape (num_res, 14) containing the atom37 indices for this protein lowercase__ = restype_atomaa_to_atomaa[protein_aatype] lowercase__ = restype_atomaa_mask[protein_aatype] lowercase__ = residx_atomaa_mask lowercase__ = residx_atomaa_to_atomaa.long() # create the gather indices for mapping back lowercase__ = restype_atomaa_to_atomaa[protein_aatype] lowercase__ = residx_atomaa_to_atomaa.long() # create the corresponding mask lowercase__ = torch.zeros([21, 37] , dtype=torch.floataa , device=protein['''aatype'''].device ) for restype, restype_letter in enumerate(rc.restypes ): lowercase__ = rc.restype_atoa[restype_letter] lowercase__ = rc.residue_atoms[restype_name] for atom_name in atom_names: lowercase__ = rc.atom_order[atom_name] lowercase__ = 1 lowercase__ = restype_atomaa_mask[protein_aatype] lowercase__ = residx_atomaa_mask return protein def a ( lowerCamelCase_ ): '''simple docstring''' lowercase__ = tree_map(lambda lowerCamelCase_ : torch.tensor(lowerCamelCase_ , device=batch['''aatype'''].device ) , lowerCamelCase_ , np.ndarray ) lowercase__ = tensor_tree_map(lambda lowerCamelCase_ : np.array(lowerCamelCase_ ) , make_atomaa_masks(lowerCamelCase_ ) ) return out
207
0
"""simple docstring""" import math def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : int ): assert isinstance(_UpperCAmelCase , _UpperCAmelCase ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or not number % 2: # Negatives, 0, 1 and all even numbers are not primes return False lowerCAmelCase = range(3 , int(math.sqrt(_UpperCAmelCase ) + 1 ) , 2 ) return not any(not number % i for i in odd_numbers ) def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : Dict , _UpperCAmelCase : List[str]=1 , **_UpperCAmelCase : Tuple ): lowerCAmelCase = factor * value lowerCAmelCase = value while not is_prime(_UpperCAmelCase ): value += 1 if not ("desc" in kwargs and kwargs["desc"] is True) else -1 if value == first_value_val: return next_prime(value + 1 , **_UpperCAmelCase ) return value
309
"""simple docstring""" from __future__ import annotations def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : str , _UpperCAmelCase : list[str] | None = None ): lowerCAmelCase = word_bank or [] # create a table lowerCAmelCase = len(_UpperCAmelCase ) + 1 lowerCAmelCase = [] for _ in range(_UpperCAmelCase ): table.append([] ) # seed value lowerCAmelCase = [[]] # because empty string has empty combination # iterate through the indices for i in range(_UpperCAmelCase ): # condition if table[i] != []: for word in word_bank: # slice condition if target[i : i + len(_UpperCAmelCase )] == word: lowerCAmelCase = [ [word, *way] for way in table[i] ] # adds the word to every combination the current position holds # now,push that combination to the table[i+len(word)] table[i + len(_UpperCAmelCase )] += new_combinations # combinations are in reverse order so reverse for better output for combination in table[len(_UpperCAmelCase )]: combination.reverse() return table[len(_UpperCAmelCase )] if __name__ == "__main__": print(all_construct('''jwajalapa''', ['''jwa''', '''j''', '''w''', '''a''', '''la''', '''lapa'''])) print(all_construct('''rajamati''', ['''s''', '''raj''', '''amat''', '''raja''', '''ma''', '''i''', '''t'''])) print( all_construct( '''hexagonosaurus''', ['''h''', '''ex''', '''hex''', '''ag''', '''ago''', '''ru''', '''auru''', '''rus''', '''go''', '''no''', '''o''', '''s'''], ) )
309
1
from ..utils import DummyObject, requires_backends class __lowercase ( metaclass=UpperCAmelCase_ ): """simple docstring""" _UpperCAmelCase : Optional[Any] = ['''torch''', '''transformers''', '''onnx'''] def __init__( self : Dict , *lowerCAmelCase__ : List[str] , **lowerCAmelCase__ : List[Any]): requires_backends(self , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls : List[str] , *lowerCAmelCase__ : List[Any] , **lowerCAmelCase__ : str): requires_backends(cls , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls : List[str] , *lowerCAmelCase__ : Optional[int] , **lowerCAmelCase__ : Union[str, Any]): requires_backends(cls , ["torch", "transformers", "onnx"]) class __lowercase ( metaclass=UpperCAmelCase_ ): """simple docstring""" _UpperCAmelCase : Dict = ['''torch''', '''transformers''', '''onnx'''] def __init__( self : int , *lowerCAmelCase__ : str , **lowerCAmelCase__ : Dict): requires_backends(self , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls : Optional[int] , *lowerCAmelCase__ : Any , **lowerCAmelCase__ : Optional[Any]): requires_backends(cls , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls : Union[str, Any] , *lowerCAmelCase__ : int , **lowerCAmelCase__ : List[str]): requires_backends(cls , ["torch", "transformers", "onnx"]) class __lowercase ( metaclass=UpperCAmelCase_ ): """simple docstring""" _UpperCAmelCase : Union[str, Any] = ['''torch''', '''transformers''', '''onnx'''] def __init__( self : List[Any] , *lowerCAmelCase__ : Dict , **lowerCAmelCase__ : List[Any]): requires_backends(self , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls : Optional[int] , *lowerCAmelCase__ : int , **lowerCAmelCase__ : Optional[Any]): requires_backends(cls , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls : Union[str, Any] , *lowerCAmelCase__ : List[Any] , **lowerCAmelCase__ : List[Any]): requires_backends(cls , ["torch", "transformers", "onnx"]) class __lowercase ( metaclass=UpperCAmelCase_ ): """simple docstring""" _UpperCAmelCase : int = ['''torch''', '''transformers''', '''onnx'''] def __init__( self : Union[str, Any] , *lowerCAmelCase__ : Tuple , **lowerCAmelCase__ : Optional[Any]): requires_backends(self , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls : Dict , *lowerCAmelCase__ : Union[str, Any] , **lowerCAmelCase__ : Optional[Any]): requires_backends(cls , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls : Tuple , *lowerCAmelCase__ : int , **lowerCAmelCase__ : Any): requires_backends(cls , ["torch", "transformers", "onnx"]) class __lowercase ( metaclass=UpperCAmelCase_ ): """simple docstring""" _UpperCAmelCase : Optional[Any] = ['''torch''', '''transformers''', '''onnx'''] def __init__( self : Any , *lowerCAmelCase__ : int , **lowerCAmelCase__ : Any): requires_backends(self , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls : Union[str, Any] , *lowerCAmelCase__ : List[Any] , **lowerCAmelCase__ : Tuple): requires_backends(cls , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls : Any , *lowerCAmelCase__ : Optional[Any] , **lowerCAmelCase__ : Union[str, Any]): requires_backends(cls , ["torch", "transformers", "onnx"]) class __lowercase ( metaclass=UpperCAmelCase_ ): """simple docstring""" _UpperCAmelCase : Optional[int] = ['''torch''', '''transformers''', '''onnx'''] def __init__( self : Any , *lowerCAmelCase__ : int , **lowerCAmelCase__ : Tuple): requires_backends(self , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls : Optional[int] , *lowerCAmelCase__ : Optional[int] , **lowerCAmelCase__ : List[Any]): requires_backends(cls , ["torch", "transformers", "onnx"]) @classmethod def _SCREAMING_SNAKE_CASE ( cls : Any , *lowerCAmelCase__ : str , **lowerCAmelCase__ : Union[str, Any]): requires_backends(cls , ["torch", "transformers", "onnx"])
13
import random import unittest import numpy as np import transformers from transformers import is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax if is_flax_available(): import os import jax.numpy as jnp from jax import jit from transformers import AutoTokenizer, FlaxAutoModelForCausalLM from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model lowerCAmelCase : Any = """0.12""" # assumed parallelism: 8 if is_torch_available(): import torch def A_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=None ): if rng is None: SCREAMING_SNAKE_CASE_: List[Any] = random.Random() SCREAMING_SNAKE_CASE_: Optional[Any] = 1 for dim in shape: total_dims *= dim SCREAMING_SNAKE_CASE_: Optional[Any] = [] for _ in range(_UpperCAmelCase ): values.append(rng.randint(0 , vocab_size - 1 ) ) SCREAMING_SNAKE_CASE_: List[Any] = np.array(_UpperCAmelCase , dtype=jnp.intaa ).reshape(_UpperCAmelCase ) return output def A_ ( _UpperCAmelCase , _UpperCAmelCase=None ): SCREAMING_SNAKE_CASE_: Optional[int] = ids_tensor(_UpperCAmelCase , vocab_size=2 , rng=_UpperCAmelCase ) # make sure that at least one token is attended to for each batch SCREAMING_SNAKE_CASE_: Optional[Any] = 1 return attn_mask @require_flax class __lowercase : """simple docstring""" _UpperCAmelCase : Any = None _UpperCAmelCase : List[Any] = () def _SCREAMING_SNAKE_CASE ( self : List[Any]): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Tuple = self.model_tester.prepare_config_and_inputs_for_common() # cut to half length & take max batch_size 3 SCREAMING_SNAKE_CASE_: str = 2 SCREAMING_SNAKE_CASE_: Optional[int] = inputs["input_ids"].shape[-1] // 2 SCREAMING_SNAKE_CASE_: List[str] = inputs["input_ids"][:max_batch_size, :sequence_length] SCREAMING_SNAKE_CASE_: Any = jnp.ones_like(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Optional[int] = attention_mask[:max_batch_size, :sequence_length] # generate max 5 tokens SCREAMING_SNAKE_CASE_: Optional[Any] = input_ids.shape[-1] + 5 if config.eos_token_id is not None and config.pad_token_id is None: # hack to allow generate for models such as GPT2 as is done in `generate()` SCREAMING_SNAKE_CASE_: Optional[Any] = config.eos_token_id return config, input_ids, attention_mask, max_length @is_pt_flax_cross_test def _SCREAMING_SNAKE_CASE ( self : Tuple): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Optional[int] = self._get_input_ids_and_config() SCREAMING_SNAKE_CASE_: Union[str, Any] = False SCREAMING_SNAKE_CASE_: Dict = max_length SCREAMING_SNAKE_CASE_: List[Any] = 0 for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE_: int = model_class(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Tuple = model_class.__name__[4:] # Skip the "Flax" at the beginning SCREAMING_SNAKE_CASE_: List[Any] = getattr(lowerCAmelCase__ , lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Union[str, Any] = pt_model_class(lowerCAmelCase__).eval() SCREAMING_SNAKE_CASE_: str = load_flax_weights_in_pytorch_model(lowerCAmelCase__ , flax_model.params) SCREAMING_SNAKE_CASE_: List[Any] = flax_model.generate(lowerCAmelCase__).sequences SCREAMING_SNAKE_CASE_: str = pt_model.generate(torch.tensor(lowerCAmelCase__ , dtype=torch.long)) if flax_generation_outputs.shape[-1] > pt_generation_outputs.shape[-1]: SCREAMING_SNAKE_CASE_: List[Any] = flax_generation_outputs[:, : pt_generation_outputs.shape[-1]] self.assertListEqual(pt_generation_outputs.numpy().tolist() , flax_generation_outputs.tolist()) def _SCREAMING_SNAKE_CASE ( self : Dict): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Optional[Any] = self._get_input_ids_and_config() SCREAMING_SNAKE_CASE_: Optional[int] = False SCREAMING_SNAKE_CASE_: Optional[int] = max_length for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE_: Union[str, Any] = model_class(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: str = model.generate(lowerCAmelCase__).sequences self.assertEqual(generation_outputs.shape[-1] , lowerCAmelCase__) SCREAMING_SNAKE_CASE_: List[Any] = jit(model.generate) SCREAMING_SNAKE_CASE_: Union[str, Any] = jit_generate(lowerCAmelCase__).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist()) def _SCREAMING_SNAKE_CASE ( self : List[str]): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Tuple = self._get_input_ids_and_config() SCREAMING_SNAKE_CASE_: Optional[Any] = True SCREAMING_SNAKE_CASE_: Dict = max_length for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE_: Tuple = model_class(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: int = model.generate(lowerCAmelCase__).sequences self.assertEqual(generation_outputs.shape[-1] , lowerCAmelCase__) SCREAMING_SNAKE_CASE_: List[str] = jit(model.generate) SCREAMING_SNAKE_CASE_: Dict = jit_generate(lowerCAmelCase__).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist()) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Any = self._get_input_ids_and_config() SCREAMING_SNAKE_CASE_: int = False SCREAMING_SNAKE_CASE_: Optional[int] = max_length SCREAMING_SNAKE_CASE_: Optional[int] = 2 for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE_: List[str] = model_class(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: str = model.generate(lowerCAmelCase__).sequences self.assertEqual(generation_outputs.shape[-1] , lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Any = jit(model.generate) SCREAMING_SNAKE_CASE_: Optional[int] = jit_generate(lowerCAmelCase__).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist()) def _SCREAMING_SNAKE_CASE ( self : Optional[int]): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Dict = self._get_input_ids_and_config() SCREAMING_SNAKE_CASE_: str = False SCREAMING_SNAKE_CASE_: int = max_length SCREAMING_SNAKE_CASE_: str = 2 SCREAMING_SNAKE_CASE_: Optional[Any] = 2 for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE_: str = model_class(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Optional[Any] = model.generate(lowerCAmelCase__).sequences self.assertEqual(generation_outputs.shape[0] , input_ids.shape[0] * config.num_return_sequences) def _SCREAMING_SNAKE_CASE ( self : Any): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Tuple = self._get_input_ids_and_config() SCREAMING_SNAKE_CASE_: Tuple = True SCREAMING_SNAKE_CASE_: List[str] = max_length SCREAMING_SNAKE_CASE_: Any = 0.8 SCREAMING_SNAKE_CASE_: Any = 10 SCREAMING_SNAKE_CASE_: List[str] = 0.3 SCREAMING_SNAKE_CASE_: Tuple = 1 SCREAMING_SNAKE_CASE_: Union[str, Any] = 8 SCREAMING_SNAKE_CASE_: int = 9 for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE_: List[str] = model_class(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: str = model.generate(lowerCAmelCase__).sequences self.assertEqual(generation_outputs.shape[-1] , lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Dict = jit(model.generate) SCREAMING_SNAKE_CASE_: List[Any] = jit_generate(lowerCAmelCase__).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist()) def _SCREAMING_SNAKE_CASE ( self : List[Any]): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: str = self._get_input_ids_and_config() SCREAMING_SNAKE_CASE_: Any = max_length SCREAMING_SNAKE_CASE_: int = 1 SCREAMING_SNAKE_CASE_: Union[str, Any] = 8 SCREAMING_SNAKE_CASE_: List[Any] = 9 for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE_: int = model_class(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Union[str, Any] = model.generate(lowerCAmelCase__).sequences self.assertEqual(generation_outputs.shape[-1] , lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Optional[Any] = jit(model.generate) SCREAMING_SNAKE_CASE_: List[str] = jit_generate(lowerCAmelCase__).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist()) def _SCREAMING_SNAKE_CASE ( self : str): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Dict = self._get_input_ids_and_config() SCREAMING_SNAKE_CASE_: Any = max_length SCREAMING_SNAKE_CASE_: List[str] = 2 SCREAMING_SNAKE_CASE_: str = 1 SCREAMING_SNAKE_CASE_: Tuple = 8 SCREAMING_SNAKE_CASE_: List[Any] = 9 for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE_: Optional[int] = model_class(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: str = model.generate(lowerCAmelCase__).sequences self.assertEqual(generation_outputs.shape[-1] , lowerCAmelCase__) SCREAMING_SNAKE_CASE_: int = jit(model.generate) SCREAMING_SNAKE_CASE_: List[str] = jit_generate(lowerCAmelCase__).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist()) def _SCREAMING_SNAKE_CASE ( self : str): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Tuple = self._get_input_ids_and_config() # pad attention mask on the left SCREAMING_SNAKE_CASE_: Dict = attention_mask.at[(0, 0)].set(0) SCREAMING_SNAKE_CASE_: Dict = False SCREAMING_SNAKE_CASE_: Optional[int] = max_length for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE_: Any = model_class(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: List[Any] = model.generate(lowerCAmelCase__ , attention_mask=lowerCAmelCase__).sequences self.assertEqual(generation_outputs.shape[-1] , lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Optional[int] = jit(model.generate) SCREAMING_SNAKE_CASE_: List[Any] = jit_generate(lowerCAmelCase__ , attention_mask=lowerCAmelCase__).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist()) def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Union[str, Any] = self._get_input_ids_and_config() # pad attention mask on the left SCREAMING_SNAKE_CASE_: List[Any] = attention_mask.at[(0, 0)].set(0) SCREAMING_SNAKE_CASE_: Optional[int] = True SCREAMING_SNAKE_CASE_: Union[str, Any] = max_length for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE_: str = model_class(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Dict = model.generate(lowerCAmelCase__ , attention_mask=lowerCAmelCase__).sequences self.assertEqual(generation_outputs.shape[-1] , lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Optional[int] = jit(model.generate) SCREAMING_SNAKE_CASE_: Optional[Any] = jit_generate(lowerCAmelCase__ , attention_mask=lowerCAmelCase__).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist()) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Tuple = self._get_input_ids_and_config() # pad attention mask on the left SCREAMING_SNAKE_CASE_: Dict = attention_mask.at[(0, 0)].set(0) SCREAMING_SNAKE_CASE_: Optional[Any] = 2 SCREAMING_SNAKE_CASE_: Any = max_length for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE_: Tuple = model_class(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Optional[int] = model.generate(lowerCAmelCase__ , attention_mask=lowerCAmelCase__).sequences self.assertEqual(generation_outputs.shape[-1] , lowerCAmelCase__) SCREAMING_SNAKE_CASE_: str = jit(model.generate) SCREAMING_SNAKE_CASE_: Union[str, Any] = jit_generate(lowerCAmelCase__ , attention_mask=lowerCAmelCase__).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist()) @require_flax class __lowercase ( unittest.TestCase ): """simple docstring""" def _SCREAMING_SNAKE_CASE ( self : List[Any]): SCREAMING_SNAKE_CASE_: Tuple = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-bert") SCREAMING_SNAKE_CASE_: List[Any] = FlaxAutoModelForCausalLM.from_pretrained("hf-internal-testing/tiny-bert-flax-only") SCREAMING_SNAKE_CASE_: Optional[int] = "Hello world" SCREAMING_SNAKE_CASE_: List[Any] = tokenizer(lowerCAmelCase__ , return_tensors="np").input_ids # typos are quickly detected (the correct argument is `do_sample`) with self.assertRaisesRegex(lowerCAmelCase__ , "do_samples"): model.generate(lowerCAmelCase__ , do_samples=lowerCAmelCase__) # arbitrary arguments that will not be used anywhere are also not accepted with self.assertRaisesRegex(lowerCAmelCase__ , "foo"): SCREAMING_SNAKE_CASE_: str = {"foo": "bar"} model.generate(lowerCAmelCase__ , **lowerCAmelCase__)
13
1
"""simple docstring""" import unittest from transformers import PegasusTokenizer, PegasusTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin a = get_tests_dir('''fixtures/test_sentencepiece_no_bos.model''') @require_sentencepiece @require_tokenizers class lowercase_ ( __lowerCAmelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase : Optional[int] = PegasusTokenizer UpperCAmelCase : List[str] = PegasusTokenizerFast UpperCAmelCase : Optional[int] = True UpperCAmelCase : List[str] = True def lowerCAmelCase_ ( self : List[str] ): super().setUp() # We have a SentencePiece fixture for testing _A = PegasusTokenizer(_UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def lowerCAmelCase_ ( self : Any ): return PegasusTokenizer.from_pretrained('google/pegasus-large' ) def lowerCAmelCase_ ( self : str , **_UpperCAmelCase : int ): return PegasusTokenizer.from_pretrained(self.tmpdirname , **_UpperCAmelCase ) def lowerCAmelCase_ ( self : Optional[Any] , _UpperCAmelCase : Tuple ): return ("This is a test", "This is a test") def lowerCAmelCase_ ( self : Dict ): _A = '</s>' _A = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_UpperCAmelCase ) , _UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_UpperCAmelCase ) , _UpperCAmelCase ) def lowerCAmelCase_ ( self : List[str] ): _A = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<pad>' ) self.assertEqual(vocab_keys[1] , '</s>' ) self.assertEqual(vocab_keys[-1] , 'v' ) self.assertEqual(len(_UpperCAmelCase ) , 1_103 ) def lowerCAmelCase_ ( self : List[str] ): self.assertEqual(self.get_tokenizer().vocab_size , 1_103 ) def lowerCAmelCase_ ( self : Optional[int] ): _A = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) _A = self.tokenizer_class.from_pretrained(self.tmpdirname ) _A = ( 'Let\'s see which <unk> is the better <unk_token_11> one <mask_1> It seems like this <mask_2> was important' ' </s> <pad> <pad> <pad>' ) _A = rust_tokenizer([raw_input_str] , return_tensors=_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ).input_ids[0] _A = py_tokenizer([raw_input_str] , return_tensors=_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ).input_ids[0] self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) def lowerCAmelCase_ ( self : str ): _A = self._large_tokenizer # <mask_1> masks whole sentence while <mask_2> masks single word _A = '<mask_1> To ensure a <mask_2> flow of bank resolutions.' _A = [2, 413, 615, 114, 3, 1_971, 113, 1_679, 10_710, 107, 1] _A = tokenizer([raw_input_str] , return_tensors=_UpperCAmelCase ).input_ids[0] self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) def lowerCAmelCase_ ( self : Any ): _A = self._large_tokenizer # The tracebacks for the following asserts are **better** without messages or self.assertEqual assert tokenizer.vocab_size == 96_103 assert tokenizer.pad_token_id == 0 assert tokenizer.eos_token_id == 1 assert tokenizer.offset == 103 assert tokenizer.unk_token_id == tokenizer.offset + 2 == 105 assert tokenizer.unk_token == "<unk>" assert tokenizer.model_max_length == 1_024 _A = 'To ensure a smooth flow of bank resolutions.' _A = [413, 615, 114, 2_291, 1_971, 113, 1_679, 10_710, 107, 1] _A = tokenizer([raw_input_str] , return_tensors=_UpperCAmelCase ).input_ids[0] self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) assert tokenizer.convert_ids_to_tokens([0, 1, 2, 3] ) == ["<pad>", "</s>", "<mask_1>", "<mask_2>"] @require_torch def lowerCAmelCase_ ( self : List[str] ): _A = ['This is going to be way too long.' * 150, 'short example'] _A = ['not super long but more than 5 tokens', 'tiny'] _A = self._large_tokenizer(_UpperCAmelCase , padding=_UpperCAmelCase , truncation=_UpperCAmelCase , return_tensors='pt' ) _A = self._large_tokenizer( text_target=_UpperCAmelCase , max_length=5 , padding=_UpperCAmelCase , truncation=_UpperCAmelCase , return_tensors='pt' ) assert batch.input_ids.shape == (2, 1_024) assert batch.attention_mask.shape == (2, 1_024) assert targets["input_ids"].shape == (2, 5) assert len(_UpperCAmelCase ) == 2 # input_ids, attention_mask. @slow def lowerCAmelCase_ ( self : Dict ): # fmt: off _A = {'input_ids': [[38_979, 143, 18_485, 606, 130, 26_669, 87_686, 121, 54_189, 1_129, 111, 26_669, 87_686, 121, 9_114, 14_787, 121, 13_249, 158, 592, 956, 121, 14_621, 31_576, 143, 62_613, 108, 9_688, 930, 43_430, 11_562, 62_613, 304, 108, 11_443, 897, 108, 9_314, 17_415, 63_399, 108, 11_443, 7_614, 18_316, 118, 4_284, 7_148, 12_430, 143, 1_400, 25_703, 158, 111, 4_284, 7_148, 11_772, 143, 21_297, 1_064, 158, 122, 204, 3_506, 1_754, 1_133, 14_787, 1_581, 115, 33_224, 4_482, 111, 1_355, 110, 29_173, 317, 50_833, 108, 20_147, 94_665, 111, 77_198, 107, 1], [110, 62_613, 117, 638, 112, 1_133, 121, 20_098, 1_355, 79_050, 13_872, 135, 1_596, 53_541, 1_352, 141, 13_039, 5_542, 124, 302, 518, 111, 268, 2_956, 115, 149, 4_427, 107, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [139, 1_235, 2_799, 18_289, 17_780, 204, 109, 9_474, 1_296, 107, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_UpperCAmelCase , model_name='google/bigbird-pegasus-large-arxiv' , revision='ba85d0851d708441f91440d509690f1ab6353415' , ) @require_sentencepiece @require_tokenizers class lowercase_ ( __lowerCAmelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase : Union[str, Any] = PegasusTokenizer UpperCAmelCase : Tuple = PegasusTokenizerFast UpperCAmelCase : Union[str, Any] = True UpperCAmelCase : Dict = True def lowerCAmelCase_ ( self : Optional[Any] ): super().setUp() # We have a SentencePiece fixture for testing _A = PegasusTokenizer(_UpperCAmelCase , offset=0 , mask_token_sent=_UpperCAmelCase , mask_token='[MASK]' ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def lowerCAmelCase_ ( self : int ): return PegasusTokenizer.from_pretrained('google/bigbird-pegasus-large-arxiv' ) def lowerCAmelCase_ ( self : List[Any] , **_UpperCAmelCase : Any ): return PegasusTokenizer.from_pretrained(self.tmpdirname , **_UpperCAmelCase ) def lowerCAmelCase_ ( self : List[Any] , _UpperCAmelCase : Optional[int] ): return ("This is a test", "This is a test") def lowerCAmelCase_ ( self : Optional[int] ): _A = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) _A = self.tokenizer_class.from_pretrained(self.tmpdirname ) _A = ( 'Let\'s see which <unk> is the better <unk_token> one [MASK] It seems like this [MASK] was important </s>' ' <pad> <pad> <pad>' ) _A = rust_tokenizer([raw_input_str] , return_tensors=_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ).input_ids[0] _A = py_tokenizer([raw_input_str] , return_tensors=_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ).input_ids[0] self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) @require_torch def lowerCAmelCase_ ( self : Optional[int] ): _A = ['This is going to be way too long.' * 1_000, 'short example'] _A = ['not super long but more than 5 tokens', 'tiny'] _A = self._large_tokenizer(_UpperCAmelCase , padding=_UpperCAmelCase , truncation=_UpperCAmelCase , return_tensors='pt' ) _A = self._large_tokenizer( text_target=_UpperCAmelCase , max_length=5 , padding=_UpperCAmelCase , truncation=_UpperCAmelCase , return_tensors='pt' ) assert batch.input_ids.shape == (2, 4_096) assert batch.attention_mask.shape == (2, 4_096) assert targets["input_ids"].shape == (2, 5) assert len(_UpperCAmelCase ) == 2 # input_ids, attention_mask. def lowerCAmelCase_ ( self : Optional[Any] ): _A = ( 'This is an example string that is used to test the original TF implementation against the HF' ' implementation' ) _A = self._large_tokenizer(_UpperCAmelCase ).input_ids self.assertListEqual( _UpperCAmelCase , [182, 117, 142, 587, 4_211, 120, 117, 263, 112, 804, 109, 856, 25_016, 3_137, 464, 109, 26_955, 3_137, 1] , )
361
"""simple docstring""" import logging import os from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union from filelock import FileLock from transformers import PreTrainedTokenizer, is_tf_available, is_torch_available a = logging.getLogger(__name__) @dataclass class lowercase_ : '''simple docstring''' UpperCAmelCase : str UpperCAmelCase : List[str] UpperCAmelCase : Optional[List[str]] @dataclass class lowercase_ : '''simple docstring''' UpperCAmelCase : List[int] UpperCAmelCase : List[int] UpperCAmelCase : Optional[List[int]] = None UpperCAmelCase : Optional[List[int]] = None class lowercase_ ( __lowerCAmelCase ): '''simple docstring''' UpperCAmelCase : Any = '''train''' UpperCAmelCase : Tuple = '''dev''' UpperCAmelCase : int = '''test''' class lowercase_ : '''simple docstring''' @staticmethod def lowerCAmelCase_ ( _UpperCAmelCase : int , _UpperCAmelCase : Union[Split, str] ): raise NotImplementedError @staticmethod def lowerCAmelCase_ ( _UpperCAmelCase : str ): raise NotImplementedError @staticmethod def lowerCAmelCase_ ( _UpperCAmelCase : List[InputExample] , _UpperCAmelCase : List[str] , _UpperCAmelCase : int , _UpperCAmelCase : PreTrainedTokenizer , _UpperCAmelCase : List[str]=False , _UpperCAmelCase : List[str]="[CLS]" , _UpperCAmelCase : List[Any]=1 , _UpperCAmelCase : Tuple="[SEP]" , _UpperCAmelCase : List[str]=False , _UpperCAmelCase : Optional[Any]=False , _UpperCAmelCase : str=0 , _UpperCAmelCase : Optional[int]=0 , _UpperCAmelCase : Any=-100 , _UpperCAmelCase : Union[str, Any]=0 , _UpperCAmelCase : List[Any]=True , ): _A = {label: i for i, label in enumerate(_UpperCAmelCase )} _A = [] for ex_index, example in enumerate(_UpperCAmelCase ): if ex_index % 10_000 == 0: logger.info('Writing example %d of %d' , _UpperCAmelCase , len(_UpperCAmelCase ) ) _A = [] _A = [] for word, label in zip(example.words , example.labels ): _A = tokenizer.tokenize(_UpperCAmelCase ) # bert-base-multilingual-cased sometimes output "nothing ([]) when calling tokenize with just a space. if len(_UpperCAmelCase ) > 0: tokens.extend(_UpperCAmelCase ) # Use the real label id for the first token of the word, and padding ids for the remaining tokens label_ids.extend([label_map[label]] + [pad_token_label_id] * (len(_UpperCAmelCase ) - 1) ) # Account for [CLS] and [SEP] with "- 2" and with "- 3" for RoBERTa. _A = tokenizer.num_special_tokens_to_add() if len(_UpperCAmelCase ) > max_seq_length - special_tokens_count: _A = tokens[: (max_seq_length - special_tokens_count)] _A = label_ids[: (max_seq_length - special_tokens_count)] # The convention in BERT is: # (a) For sequence pairs: # tokens: [CLS] is this jack ##son ##ville ? [SEP] no it is not . [SEP] # type_ids: 0 0 0 0 0 0 0 0 1 1 1 1 1 1 # (b) For single sequences: # tokens: [CLS] the dog is hairy . [SEP] # type_ids: 0 0 0 0 0 0 0 # # Where "type_ids" are used to indicate whether this is the first # sequence or the second sequence. The embedding vectors for `type=0` and # `type=1` were learned during pre-training and are added to the wordpiece # embedding vector (and position vector). This is not *strictly* necessary # since the [SEP] token unambiguously separates the sequences, but it makes # it easier for the model to learn the concept of sequences. # # For classification tasks, the first vector (corresponding to [CLS]) is # used as the "sentence vector". Note that this only makes sense because # the entire model is fine-tuned. tokens += [sep_token] label_ids += [pad_token_label_id] if sep_token_extra: # roberta uses an extra separator b/w pairs of sentences tokens += [sep_token] label_ids += [pad_token_label_id] _A = [sequence_a_segment_id] * len(_UpperCAmelCase ) if cls_token_at_end: tokens += [cls_token] label_ids += [pad_token_label_id] segment_ids += [cls_token_segment_id] else: _A = [cls_token] + tokens _A = [pad_token_label_id] + label_ids _A = [cls_token_segment_id] + segment_ids _A = tokenizer.convert_tokens_to_ids(_UpperCAmelCase ) # The mask has 1 for real tokens and 0 for padding tokens. Only real # tokens are attended to. _A = [1 if mask_padding_with_zero else 0] * len(_UpperCAmelCase ) # Zero-pad up to the sequence length. _A = max_seq_length - len(_UpperCAmelCase ) if pad_on_left: _A = ([pad_token] * padding_length) + input_ids _A = ([0 if mask_padding_with_zero else 1] * padding_length) + input_mask _A = ([pad_token_segment_id] * padding_length) + segment_ids _A = ([pad_token_label_id] * padding_length) + label_ids else: input_ids += [pad_token] * padding_length input_mask += [0 if mask_padding_with_zero else 1] * padding_length segment_ids += [pad_token_segment_id] * padding_length label_ids += [pad_token_label_id] * padding_length assert len(_UpperCAmelCase ) == max_seq_length assert len(_UpperCAmelCase ) == max_seq_length assert len(_UpperCAmelCase ) == max_seq_length assert len(_UpperCAmelCase ) == max_seq_length if ex_index < 5: logger.info('*** Example ***' ) logger.info('guid: %s' , example.guid ) logger.info('tokens: %s' , ' '.join([str(_UpperCAmelCase ) for x in tokens] ) ) logger.info('input_ids: %s' , ' '.join([str(_UpperCAmelCase ) for x in input_ids] ) ) logger.info('input_mask: %s' , ' '.join([str(_UpperCAmelCase ) for x in input_mask] ) ) logger.info('segment_ids: %s' , ' '.join([str(_UpperCAmelCase ) for x in segment_ids] ) ) logger.info('label_ids: %s' , ' '.join([str(_UpperCAmelCase ) for x in label_ids] ) ) if "token_type_ids" not in tokenizer.model_input_names: _A = None features.append( InputFeatures( input_ids=_UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase , label_ids=_UpperCAmelCase ) ) return features if is_torch_available(): import torch from torch import nn from torch.utils.data import Dataset class lowercase_ ( __lowerCAmelCase ): '''simple docstring''' UpperCAmelCase : List[InputFeatures] UpperCAmelCase : int = nn.CrossEntropyLoss().ignore_index def __init__( self : int , _UpperCAmelCase : TokenClassificationTask , _UpperCAmelCase : str , _UpperCAmelCase : PreTrainedTokenizer , _UpperCAmelCase : List[str] , _UpperCAmelCase : str , _UpperCAmelCase : Optional[int] = None , _UpperCAmelCase : Optional[Any]=False , _UpperCAmelCase : Split = Split.train , ): # Load data features from cache or dataset file _A = os.path.join( _UpperCAmelCase , 'cached_{}_{}_{}'.format(mode.value , tokenizer.__class__.__name__ , str(_UpperCAmelCase ) ) , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. _A = cached_features_file + '.lock' with FileLock(_UpperCAmelCase ): if os.path.exists(_UpperCAmelCase ) and not overwrite_cache: logger.info(F'''Loading features from cached file {cached_features_file}''' ) _A = torch.load(_UpperCAmelCase ) else: logger.info(F'''Creating features from dataset file at {data_dir}''' ) _A = token_classification_task.read_examples_from_file(_UpperCAmelCase , _UpperCAmelCase ) # TODO clean up all this to leverage built-in features of tokenizers _A = token_classification_task.convert_examples_to_features( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , cls_token_at_end=bool(model_type in ['xlnet'] ) , cls_token=tokenizer.cls_token , cls_token_segment_id=2 if model_type in ['xlnet'] else 0 , sep_token=tokenizer.sep_token , sep_token_extra=_UpperCAmelCase , pad_on_left=bool(tokenizer.padding_side == 'left' ) , pad_token=tokenizer.pad_token_id , pad_token_segment_id=tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) logger.info(F'''Saving features into cached file {cached_features_file}''' ) torch.save(self.features , _UpperCAmelCase ) def __len__( self : Dict ): return len(self.features ) def __getitem__( self : int , _UpperCAmelCase : Union[str, Any] ): return self.features[i] if is_tf_available(): import tensorflow as tf class lowercase_ : '''simple docstring''' UpperCAmelCase : List[InputFeatures] UpperCAmelCase : int = -100 def __init__( self : int , _UpperCAmelCase : TokenClassificationTask , _UpperCAmelCase : str , _UpperCAmelCase : PreTrainedTokenizer , _UpperCAmelCase : List[str] , _UpperCAmelCase : str , _UpperCAmelCase : Optional[int] = None , _UpperCAmelCase : Optional[Any]=False , _UpperCAmelCase : Split = Split.train , ): _A = token_classification_task.read_examples_from_file(_UpperCAmelCase , _UpperCAmelCase ) # TODO clean up all this to leverage built-in features of tokenizers _A = token_classification_task.convert_examples_to_features( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , cls_token_at_end=bool(model_type in ['xlnet'] ) , cls_token=tokenizer.cls_token , cls_token_segment_id=2 if model_type in ['xlnet'] else 0 , sep_token=tokenizer.sep_token , sep_token_extra=_UpperCAmelCase , pad_on_left=bool(tokenizer.padding_side == 'left' ) , pad_token=tokenizer.pad_token_id , pad_token_segment_id=tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) def gen(): for ex in self.features: if ex.token_type_ids is None: yield ( {"input_ids": ex.input_ids, "attention_mask": ex.attention_mask}, ex.label_ids, ) else: yield ( { "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label_ids, ) if "token_type_ids" not in tokenizer.model_input_names: _A = tf.data.Dataset.from_generator( _UpperCAmelCase , ({'input_ids': tf.intaa, 'attention_mask': tf.intaa}, tf.intaa) , ( {'input_ids': tf.TensorShape([None] ), 'attention_mask': tf.TensorShape([None] )}, tf.TensorShape([None] ), ) , ) else: _A = tf.data.Dataset.from_generator( _UpperCAmelCase , ({'input_ids': tf.intaa, 'attention_mask': tf.intaa, 'token_type_ids': tf.intaa}, tf.intaa) , ( { 'input_ids': tf.TensorShape([None] ), 'attention_mask': tf.TensorShape([None] ), 'token_type_ids': tf.TensorShape([None] ), }, tf.TensorShape([None] ), ) , ) def lowerCAmelCase_ ( self : Dict ): _A = self.dataset.apply(tf.data.experimental.assert_cardinality(len(self.features ) ) ) return self.dataset def __len__( self : Tuple ): return len(self.features ) def __getitem__( self : Dict , _UpperCAmelCase : Optional[int] ): return self.features[i]
271
0
"""simple docstring""" # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. _a = abspath(join(dirname(dirname(dirname(__file__))), 'src')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='ignore', category=FutureWarning) def _A ( UpperCamelCase_ : Union[str, Any]) -> Union[str, Any]: '''simple docstring''' from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(__a) def _A ( UpperCamelCase_ : List[Any]) -> Any: '''simple docstring''' from transformers.testing_utils import pytest_terminal_summary_main __lowercase = terminalreporter.config.getoption("--make-reports") if make_reports: pytest_terminal_summary_main(__a, id=__a)
17
import re import jax.numpy as jnp from flax.traverse_util import flatten_dict, unflatten_dict from jax.random import PRNGKey from ..utils import logging a__ = logging.get_logger(__name__) def __UpperCAmelCase ( __a : Dict ) -> Tuple: """simple docstring""" _a : Any = R'''\w+[.]\d+''' _a : Union[str, Any] = re.findall(__a ,__a ) for pat in pats: _a : int = key.replace(__a ,'''_'''.join(pat.split('''.''' ) ) ) return key def __UpperCAmelCase ( __a : List[str] ,__a : Union[str, Any] ,__a : Optional[int] ) -> Tuple: """simple docstring""" _a : Dict = pt_tuple_key[:-1] + ('''scale''',) if ( any('''norm''' in str_ for str_ in pt_tuple_key ) and (pt_tuple_key[-1] == "bias") and (pt_tuple_key[:-1] + ("bias",) not in random_flax_state_dict) and (pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict) ): _a : Dict = pt_tuple_key[:-1] + ('''scale''',) return renamed_pt_tuple_key, pt_tensor elif pt_tuple_key[-1] in ["weight", "gamma"] and pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict: _a : Optional[int] = pt_tuple_key[:-1] + ('''scale''',) return renamed_pt_tuple_key, pt_tensor # embedding if pt_tuple_key[-1] == "weight" and pt_tuple_key[:-1] + ("embedding",) in random_flax_state_dict: _a : Union[str, Any] = pt_tuple_key[:-1] + ('''embedding''',) return renamed_pt_tuple_key, pt_tensor # conv layer _a : Tuple = pt_tuple_key[:-1] + ('''kernel''',) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4: _a : List[str] = pt_tensor.transpose(2 ,3 ,1 ,0 ) return renamed_pt_tuple_key, pt_tensor # linear layer _a : Dict = pt_tuple_key[:-1] + ('''kernel''',) if pt_tuple_key[-1] == "weight": _a : Union[str, Any] = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight _a : Dict = pt_tuple_key[:-1] + ('''weight''',) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias _a : Union[str, Any] = pt_tuple_key[:-1] + ('''bias''',) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def __UpperCAmelCase ( __a : Dict ,__a : str ,__a : str=42 ) -> Optional[int]: """simple docstring""" _a : Tuple = {k: v.numpy() for k, v in pt_state_dict.items()} # Step 2: Since the model is stateless, get random Flax params _a : List[Any] = flax_model.init_weights(PRNGKey(__a ) ) _a : Optional[int] = flatten_dict(__a ) _a : Tuple = {} # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): _a : List[str] = rename_key(__a ) _a : Optional[Any] = tuple(renamed_pt_key.split('''.''' ) ) # Correctly rename weight parameters _a , _a : List[str] = rename_key_and_reshape_tensor(__a ,__a ,__a ) if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( F"""PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape """ F"""{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) # also add unexpected weight so that warning is thrown _a : Dict = jnp.asarray(__a ) return unflatten_dict(__a )
235
0
from .data_collator import ( DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForSeqaSeq, DataCollatorForSOP, DataCollatorForTokenClassification, DataCollatorForWholeWordMask, DataCollatorWithPadding, DefaultDataCollator, default_data_collator, ) from .metrics import glue_compute_metrics, xnli_compute_metrics from .processors import ( DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor, SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels, squad_convert_examples_to_features, xnli_output_modes, xnli_processors, xnli_tasks_num_labels, )
359
import warnings from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging lowerCamelCase : Optional[Any] = logging.get_logger(__name__) lowerCamelCase : Optional[int] = { 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/config.json', # See all BART models at https://huggingface.co/models?filter=bart } class __lowercase (UpperCamelCase__ ): """simple docstring""" _snake_case = """bart""" _snake_case = ["""past_key_values"""] _snake_case = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self , A=5_0_2_6_5 , A=1_0_2_4 , A=1_2 , A=4_0_9_6 , A=1_6 , A=1_2 , A=4_0_9_6 , A=1_6 , A=0.0 , A=0.0 , A="gelu" , A=1_0_2_4 , A=0.1 , A=0.0 , A=0.0 , A=0.02 , A=0.0 , A=False , A=True , A=3 , A=1 , A=0 , A=2 , A=True , A=2 , A=2 , **A , ) -> Any: snake_case : Optional[int] = vocab_size snake_case : Union[str, Any] = max_position_embeddings snake_case : List[str] = d_model snake_case : List[Any] = encoder_ffn_dim snake_case : Optional[Any] = encoder_layers snake_case : Union[str, Any] = encoder_attention_heads snake_case : str = decoder_ffn_dim snake_case : Union[str, Any] = decoder_layers snake_case : Any = decoder_attention_heads snake_case : Union[str, Any] = dropout snake_case : List[str] = attention_dropout snake_case : List[Any] = activation_dropout snake_case : Optional[int] = activation_function snake_case : Union[str, Any] = init_std snake_case : List[str] = encoder_layerdrop snake_case : int = decoder_layerdrop snake_case : str = classifier_dropout snake_case : List[str] = use_cache snake_case : Tuple = encoder_layers snake_case : int = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( num_labels=A , pad_token_id=A , bos_token_id=A , eos_token_id=A , is_encoder_decoder=A , decoder_start_token_id=A , forced_eos_token_id=A , **A , ) # ensure backward compatibility for BART CNN models if self.forced_bos_token_id is None and kwargs.get("""force_bos_token_to_be_generated""" , A ): snake_case : Any = self.bos_token_id warnings.warn( f"""Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. """ """The config can simply be saved and uploaded again to be fixed.""" ) class __lowercase (UpperCamelCase__ ): """simple docstring""" @property def UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: snake_case : Optional[Any] = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ] ) if self.use_past: snake_case : Tuple = {0: """batch"""} snake_case : List[Any] = {0: """batch""", 1: """past_decoder_sequence + sequence"""} else: snake_case : Union[str, Any] = {0: """batch""", 1: """decoder_sequence"""} snake_case : Any = {0: """batch""", 1: """decoder_sequence"""} if self.use_past: self.fill_with_past_key_values_(A , direction="""inputs""" ) elif self.task == "causal-lm": # TODO: figure this case out. snake_case : Union[str, Any] = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ] ) if self.use_past: snake_case , snake_case : List[Any] = self.num_layers for i in range(A ): snake_case : List[Any] = {0: """batch""", 2: """past_sequence + sequence"""} snake_case : Optional[int] = {0: """batch""", 2: """past_sequence + sequence"""} else: snake_case : Union[str, Any] = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ("""decoder_input_ids""", {0: """batch""", 1: """decoder_sequence"""}), ("""decoder_attention_mask""", {0: """batch""", 1: """decoder_sequence"""}), ] ) return common_inputs @property def UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: snake_case : Any = super().outputs else: snake_case : Any = super(A , self ).outputs if self.use_past: snake_case , snake_case : Any = self.num_layers for i in range(A ): snake_case : Any = {0: """batch""", 2: """past_sequence + sequence"""} snake_case : Union[str, Any] = {0: """batch""", 2: """past_sequence + sequence"""} return common_outputs def UpperCAmelCase ( self , A , A = -1 , A = -1 , A = False , A = None , ) -> Mapping[str, Any]: snake_case : Optional[Any] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( A , A , A , A , A ) # Generate decoder inputs snake_case : Any = seq_length if not self.use_past else 1 snake_case : List[str] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( A , A , A , A , A ) snake_case : Optional[int] = {f"""decoder_{name}""": tensor for name, tensor in decoder_inputs.items()} snake_case : List[str] = dict(**A , **A ) if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch snake_case , snake_case : Optional[int] = common_inputs["""input_ids"""].shape snake_case : Any = common_inputs["""decoder_input_ids"""].shape[1] snake_case , snake_case : Optional[Any] = self.num_attention_heads snake_case : Optional[int] = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) snake_case : Any = decoder_seq_length + 3 snake_case : List[Any] = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) snake_case : str = torch.cat( [common_inputs["""decoder_attention_mask"""], torch.ones(A , A )] , dim=1 ) snake_case : str = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered snake_case , snake_case : Any = self.num_layers snake_case : List[str] = min(A , A ) snake_case : Dict = max(A , A ) - min_num_layers snake_case : List[str] = """encoder""" if num_encoder_layers > num_decoder_layers else """decoder""" for _ in range(A ): common_inputs["past_key_values"].append( ( torch.zeros(A ), torch.zeros(A ), torch.zeros(A ), torch.zeros(A ), ) ) # TODO: test this. snake_case : Tuple = encoder_shape if remaining_side_name == """encoder""" else decoder_shape for _ in range(A , A ): common_inputs["past_key_values"].append((torch.zeros(A ), torch.zeros(A )) ) return common_inputs def UpperCAmelCase ( self , A , A = -1 , A = -1 , A = False , A = None , ) -> Mapping[str, Any]: snake_case : str = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( A , A , A , A , A ) if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch snake_case , snake_case : str = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values snake_case : Optional[int] = seqlen + 2 snake_case , snake_case : Tuple = self.num_layers snake_case , snake_case : Optional[Any] = self.num_attention_heads snake_case : Union[str, Any] = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) snake_case : Optional[Any] = common_inputs["""attention_mask"""].dtype snake_case : int = torch.cat( [common_inputs["""attention_mask"""], torch.ones(A , A , dtype=A )] , dim=1 ) snake_case : Union[str, Any] = [ (torch.zeros(A ), torch.zeros(A )) for _ in range(A ) ] return common_inputs def UpperCAmelCase ( self , A , A = -1 , A = -1 , A = False , A = None , ) -> Mapping[str, Any]: # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX snake_case : int = compute_effective_axis_dimension( A , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX snake_case : int = tokenizer.num_special_tokens_to_add(A ) snake_case : Tuple = compute_effective_axis_dimension( A , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=A ) # Generate dummy inputs according to compute batch and sequence snake_case : int = [""" """.join([tokenizer.unk_token] ) * seq_length] * batch_size snake_case : str = dict(tokenizer(A , return_tensors=A ) ) return common_inputs def UpperCAmelCase ( self , A , A = -1 , A = -1 , A = False , A = None , ) -> Mapping[str, Any]: if self.task in ["default", "seq2seq-lm"]: snake_case : Optional[Any] = self._generate_dummy_inputs_for_default_and_seqaseq_lm( A , batch_size=A , seq_length=A , is_pair=A , framework=A ) elif self.task == "causal-lm": snake_case : Optional[int] = self._generate_dummy_inputs_for_causal_lm( A , batch_size=A , seq_length=A , is_pair=A , framework=A ) else: snake_case : List[Any] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( A , batch_size=A , seq_length=A , is_pair=A , framework=A ) return common_inputs def UpperCAmelCase ( self , A , A , A , A ) -> Union[str, Any]: if self.task in ["default", "seq2seq-lm"]: snake_case : Optional[Any] = super()._flatten_past_key_values_(A , A , A , A ) else: snake_case : Union[str, Any] = super(A , self )._flatten_past_key_values_( A , A , A , A )
176
0
"""simple docstring""" from __future__ import annotations def lowercase__( __SCREAMING_SNAKE_CASE : list[int] ): # This function is recursive lowercase_ : Any = len(__SCREAMING_SNAKE_CASE ) # If the array contains only one element, we return it (it's the stop condition of # recursion) if array_length <= 1: return array # Else lowercase_ : Optional[Any] = array[0] lowercase_ : str = False lowercase_ : Dict = 1 lowercase_ : list[int] = [] while not is_found and i < array_length: if array[i] < pivot: lowercase_ : Dict = True lowercase_ : Dict = [element for element in array[i:] if element >= array[i]] lowercase_ : str = longest_subsequence(__SCREAMING_SNAKE_CASE ) if len(__SCREAMING_SNAKE_CASE ) > len(__SCREAMING_SNAKE_CASE ): lowercase_ : Dict = temp_array else: i += 1 lowercase_ : str = [element for element in array[1:] if element >= pivot] lowercase_ : Dict = [pivot, *longest_subsequence(__SCREAMING_SNAKE_CASE )] if len(__SCREAMING_SNAKE_CASE ) > len(__SCREAMING_SNAKE_CASE ): return temp_array else: return longest_subseq if __name__ == "__main__": import doctest doctest.testmod()
213
"""simple docstring""" from collections import Counter from pathlib import Path from typing import Optional, Tuple import yaml class UpperCamelCase ( yaml.SafeLoader ): def _UpperCAmelCase ( self ,__UpperCamelCase ) -> Optional[int]: '''simple docstring''' lowercase_ : str = [self.constructed_objects[key_node] for key_node, _ in node.value] lowercase_ : str = [tuple(__UpperCamelCase ) if isinstance(__UpperCamelCase ,__UpperCamelCase ) else key for key in keys] lowercase_ : List[Any] = Counter(__UpperCamelCase ) lowercase_ : str = [key for key in counter if counter[key] > 1] if duplicate_keys: raise TypeError(f'''Got duplicate yaml keys: {duplicate_keys}''' ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase=False ) -> List[Any]: '''simple docstring''' lowercase_ : Optional[int] = super().construct_mapping(__UpperCamelCase ,deep=__UpperCamelCase ) self._check_no_duplicates_on_constructed_node(__UpperCamelCase ) return mapping def lowercase__( __SCREAMING_SNAKE_CASE : str ): lowercase_ : Tuple = list(readme_content.splitlines() ) if full_content and full_content[0] == "---" and "---" in full_content[1:]: lowercase_ : Dict = full_content[1:].index('---' ) + 1 lowercase_ : Optional[int] = '\n'.join(full_content[1:sep_idx] ) return yamlblock, "\n".join(full_content[sep_idx + 1 :] ) return None, "\n".join(__SCREAMING_SNAKE_CASE ) class UpperCamelCase ( lowercase_ ): # class attributes lowercase = {'train_eval_index'} # train-eval-index in the YAML metadata @classmethod def _UpperCAmelCase ( cls ,__UpperCamelCase ) -> "DatasetMetadata": '''simple docstring''' with open(__UpperCamelCase ,encoding='utf-8' ) as readme_file: lowercase_ , lowercase_ : Optional[int] = _split_yaml_from_readme(readme_file.read() ) if yaml_string is not None: return cls.from_yaml_string(__UpperCamelCase ) else: return cls() def _UpperCAmelCase ( self ,__UpperCamelCase ) -> List[Any]: '''simple docstring''' if path.exists(): with open(__UpperCamelCase ,encoding='utf-8' ) as readme_file: lowercase_ : Dict = readme_file.read() else: lowercase_ : int = None lowercase_ : Any = self._to_readme(__UpperCamelCase ) with open(__UpperCamelCase ,'w' ,encoding='utf-8' ) as readme_file: readme_file.write(__UpperCamelCase ) def _UpperCAmelCase ( self ,__UpperCamelCase = None ) -> str: '''simple docstring''' if readme_content is not None: lowercase_ , lowercase_ : Optional[Any] = _split_yaml_from_readme(__UpperCamelCase ) lowercase_ : Optional[Any] = '---\n' + self.to_yaml_string() + '---\n' + content else: lowercase_ : Tuple = '---\n' + self.to_yaml_string() + '---\n' return full_content @classmethod def _UpperCAmelCase ( cls ,__UpperCamelCase ) -> "DatasetMetadata": '''simple docstring''' lowercase_ : List[str] = yaml.load(__UpperCamelCase ,Loader=_NoDuplicateSafeLoader ) or {} # Convert the YAML keys to DatasetMetadata fields lowercase_ : Dict = { (key.replace('-' ,'_' ) if key.replace('-' ,'_' ) in cls._FIELDS_WITH_DASHES else key): value for key, value in metadata_dict.items() } return cls(**__UpperCamelCase ) def _UpperCAmelCase ( self ) -> str: '''simple docstring''' return yaml.safe_dump( { (key.replace('_' ,'-' ) if key in self._FIELDS_WITH_DASHES else key): value for key, value in self.items() } ,sort_keys=__UpperCamelCase ,allow_unicode=__UpperCamelCase ,encoding='utf-8' ,).decode('utf-8' ) __SCREAMING_SNAKE_CASE ={ "image-classification": [], "translation": [], "image-segmentation": [], "fill-mask": [], "automatic-speech-recognition": [], "token-classification": [], "sentence-similarity": [], "audio-classification": [], "question-answering": [], "summarization": [], "zero-shot-classification": [], "table-to-text": [], "feature-extraction": [], "other": [], "multiple-choice": [], "text-classification": [], "text-to-image": [], "text2text-generation": [], "zero-shot-image-classification": [], "tabular-classification": [], "tabular-regression": [], "image-to-image": [], "tabular-to-text": [], "unconditional-image-generation": [], "text-retrieval": [], "text-to-speech": [], "object-detection": [], "audio-to-audio": [], "text-generation": [], "conversational": [], "table-question-answering": [], "visual-question-answering": [], "image-to-text": [], "reinforcement-learning": [], "voice-activity-detection": [], "time-series-forecasting": [], "document-question-answering": [], } if __name__ == "__main__": from argparse import ArgumentParser __SCREAMING_SNAKE_CASE =ArgumentParser(usage="Validate the yaml metadata block of a README.md file.") ap.add_argument("readme_filepath") __SCREAMING_SNAKE_CASE =ap.parse_args() __SCREAMING_SNAKE_CASE =Path(args.readme_filepath) __SCREAMING_SNAKE_CASE =DatasetMetadata.from_readme(readme_filepath) print(dataset_metadata) dataset_metadata.to_readme(readme_filepath)
213
1
from __future__ import annotations from collections import namedtuple from dataclasses import dataclass @dataclass class __lowerCamelCase : """simple docstring""" lowerCAmelCase__ = 42 lowerCAmelCase__ = None lowerCAmelCase__ = None SCREAMING_SNAKE_CASE__ = namedtuple("""CoinsDistribResult""", """moves excess""") def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: TreeNode | None ): '''simple docstring''' if root is None: return 0 # Validation def count_nodes(__lowerCamelCase: TreeNode | None ) -> int: if node is None: return 0 return count_nodes(node.left ) + count_nodes(node.right ) + 1 def count_coins(__lowerCamelCase: TreeNode | None ) -> int: if node is None: return 0 return count_coins(node.left ) + count_coins(node.right ) + node.data if count_nodes(__lowerCamelCase ) != count_coins(__lowerCamelCase ): raise ValueError("The nodes number should be same as the number of coins" ) # Main calculation def get_distrib(__lowerCamelCase: TreeNode | None ) -> CoinsDistribResult: if node is None: return CoinsDistribResult(0 , 1 ) lowercase_ , lowercase_ = get_distrib(node.left ) lowercase_ , lowercase_ = get_distrib(node.right ) lowercase_ = 1 - left_distrib_excess lowercase_ = 1 - right_distrib_excess lowercase_ = ( left_distrib_moves + right_distrib_moves + abs(__lowerCamelCase ) + abs(__lowerCamelCase ) ) lowercase_ = node.data - coins_to_left - coins_to_right return CoinsDistribResult(__lowerCamelCase , __lowerCamelCase ) return get_distrib(__lowerCamelCase )[0] if __name__ == "__main__": import doctest doctest.testmod()
297
def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: float ): '''simple docstring''' return 10 - x * x def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: float , __lowerCamelCase: float ): '''simple docstring''' if equation(__lowerCamelCase ) * equation(__lowerCamelCase ) >= 0: raise ValueError("Wrong space!" ) lowercase_ = a while (b - a) >= 0.01: # Find middle point lowercase_ = (a + b) / 2 # Check if middle point is root if equation(__lowerCamelCase ) == 0.0: break # Decide the side to repeat the steps if equation(__lowerCamelCase ) * equation(__lowerCamelCase ) < 0: lowercase_ = c else: lowercase_ = c return c if __name__ == "__main__": import doctest doctest.testmod() print(bisection(-2, 5)) print(bisection(0, 6))
297
1
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, XLMRobertaTokenizer from diffusers import AltDiffusionPipeline, AutoencoderKL, DDIMScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _UpperCamelCase ( snake_case_ , snake_case_ , snake_case_ , unittest.TestCase ): '''simple docstring''' _A : Union[str, Any] = AltDiffusionPipeline _A : int = TEXT_TO_IMAGE_PARAMS _A : Tuple = TEXT_TO_IMAGE_BATCH_PARAMS _A : str = TEXT_TO_IMAGE_IMAGE_PARAMS _A : List[str] = TEXT_TO_IMAGE_IMAGE_PARAMS def UpperCamelCase__ ( self : Any ): """simple docstring""" torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE : int = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=3_2 , ) __SCREAMING_SNAKE_CASE : Dict = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=lowerCAmelCase__ , set_alpha_to_one=lowerCAmelCase__ , ) torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE : List[str] = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) # TODO: address the non-deterministic text encoder (fails for save-load tests) # torch.manual_seed(0) # text_encoder_config = RobertaSeriesConfig( # hidden_size=32, # project_dim=32, # intermediate_size=37, # layer_norm_eps=1e-05, # num_attention_heads=4, # num_hidden_layers=5, # vocab_size=5002, # ) # text_encoder = RobertaSeriesModelWithTransformation(text_encoder_config) torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE : str = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , projection_dim=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5_0_0_2 , ) __SCREAMING_SNAKE_CASE : Any = CLIPTextModel(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Tuple = XLMRobertaTokenizer.from_pretrained("""hf-internal-testing/tiny-xlm-roberta""" ) __SCREAMING_SNAKE_CASE : Any = 7_7 __SCREAMING_SNAKE_CASE : List[str] = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def UpperCamelCase__ ( self : List[Any] , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : Tuple=0 ): """simple docstring""" if str(lowerCAmelCase__ ).startswith("""mps""" ): __SCREAMING_SNAKE_CASE : Optional[Any] = torch.manual_seed(lowerCAmelCase__ ) else: __SCREAMING_SNAKE_CASE : Any = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Union[str, Any] = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def UpperCamelCase__ ( self : int ): """simple docstring""" super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def UpperCamelCase__ ( self : str ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def UpperCamelCase__ ( self : Any ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[str] = """cpu""" # ensure determinism for the device-dependent torch.Generator __SCREAMING_SNAKE_CASE : int = self.get_dummy_components() torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE : Any = RobertaSeriesConfig( hidden_size=3_2 , project_dim=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=5_0_0_2 , ) # TODO: remove after fixing the non-deterministic text encoder __SCREAMING_SNAKE_CASE : List[str] = RobertaSeriesModelWithTransformation(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Optional[int] = text_encoder __SCREAMING_SNAKE_CASE : List[Any] = AltDiffusionPipeline(**lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : str = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Dict = self.get_dummy_inputs(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : str = """A photo of an astronaut""" __SCREAMING_SNAKE_CASE : int = alt_pipe(**lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Union[str, Any] = output.images __SCREAMING_SNAKE_CASE : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) __SCREAMING_SNAKE_CASE : Union[str, Any] = np.array( [0.5_74_81_62, 0.60_44_71_45, 0.48_82_12_17, 0.50_10_06_36, 0.5_43_11_85, 0.45_76_36_83, 0.49_65_76_96, 0.48_13_27_33, 0.47_57_30_93] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase__ ( self : Dict ): """simple docstring""" __SCREAMING_SNAKE_CASE : Union[str, Any] = """cpu""" # ensure determinism for the device-dependent torch.Generator __SCREAMING_SNAKE_CASE : Tuple = self.get_dummy_components() __SCREAMING_SNAKE_CASE : Tuple = PNDMScheduler(skip_prk_steps=lowerCAmelCase__ ) torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE : str = RobertaSeriesConfig( hidden_size=3_2 , project_dim=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=5_0_0_2 , ) # TODO: remove after fixing the non-deterministic text encoder __SCREAMING_SNAKE_CASE : str = RobertaSeriesModelWithTransformation(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Any = text_encoder __SCREAMING_SNAKE_CASE : str = AltDiffusionPipeline(**lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : List[Any] = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Dict = self.get_dummy_inputs(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Tuple = alt_pipe(**lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : str = output.images __SCREAMING_SNAKE_CASE : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) __SCREAMING_SNAKE_CASE : List[str] = np.array( [0.51_60_50_93, 0.5_70_72_41, 0.47_36_55_07, 0.50_57_88_86, 0.5_63_38_77, 0.4_64_25_03, 0.5_18_20_81, 0.48_76_34_84, 0.49_08_42_37] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self : Union[str, Any] ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self : List[str] ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[Any] = AltDiffusionPipeline.from_pretrained("""BAAI/AltDiffusion""" , safety_checker=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : List[Any] = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Tuple = """A painting of a squirrel eating a burger""" __SCREAMING_SNAKE_CASE : Tuple = torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE : Optional[Any] = alt_pipe([prompt] , generator=lowerCAmelCase__ , guidance_scale=6.0 , num_inference_steps=2_0 , output_type="""np""" ) __SCREAMING_SNAKE_CASE : Any = output.images __SCREAMING_SNAKE_CASE : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) __SCREAMING_SNAKE_CASE : List[str] = np.array([0.10_10, 0.08_00, 0.07_94, 0.08_85, 0.08_43, 0.07_62, 0.07_69, 0.07_29, 0.05_86] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase__ ( self : int ): """simple docstring""" __SCREAMING_SNAKE_CASE : int = DDIMScheduler.from_pretrained("""BAAI/AltDiffusion""" , subfolder="""scheduler""" ) __SCREAMING_SNAKE_CASE : int = AltDiffusionPipeline.from_pretrained("""BAAI/AltDiffusion""" , scheduler=lowerCAmelCase__ , safety_checker=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Any = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Union[str, Any] = """A painting of a squirrel eating a burger""" __SCREAMING_SNAKE_CASE : Dict = torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE : Optional[int] = alt_pipe([prompt] , generator=lowerCAmelCase__ , num_inference_steps=2 , output_type="""numpy""" ) __SCREAMING_SNAKE_CASE : Tuple = output.images __SCREAMING_SNAKE_CASE : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) __SCREAMING_SNAKE_CASE : List[Any] = np.array([0.40_19, 0.40_52, 0.38_10, 0.41_19, 0.39_16, 0.39_82, 0.46_51, 0.41_95, 0.53_23] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
112
"""simple docstring""" from __future__ import annotations from typing import Any class snake_case__ : def __init__( self , lowerCamelCase , lowerCamelCase , lowerCamelCase = 0 ): __a , __a = row, column __a = [[default_value for c in range(lowerCamelCase )] for r in range(lowerCamelCase )] def __str__( self ): __a = F"Matrix consist of {self.row} rows and {self.column} columns\n" # Make string identifier __a = 0 for row_vector in self.array: for obj in row_vector: __a = max(lowerCamelCase , len(str(lowerCamelCase ) ) ) __a = F"%{max_element_length}s" # Make string and return def single_line(lowerCamelCase ) -> str: nonlocal string_format_identifier __a = "[" line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(lowerCamelCase ) for row_vector in self.array ) return s def __repr__( self ): return str(self ) def a__ ( self , lowerCamelCase ): if not (isinstance(lowerCamelCase , (list, tuple) ) and len(lowerCamelCase ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self , lowerCamelCase ): assert self.validate_indicies(lowerCamelCase ) return self.array[loc[0]][loc[1]] def __setitem__( self , lowerCamelCase , lowerCamelCase ): assert self.validate_indicies(lowerCamelCase ) __a = value def __add__( self , lowerCamelCase ): assert isinstance(lowerCamelCase , lowerCamelCase ) assert self.row == another.row and self.column == another.column # Add __a = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): __a = self[r, c] + another[r, c] return result def __neg__( self ): __a = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): __a = -self[r, c] return result def __sub__( self , lowerCamelCase ): return self + (-another) def __mul__( self , lowerCamelCase ): if isinstance(lowerCamelCase , (int, float) ): # Scalar multiplication __a = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): __a = self[r, c] * another return result elif isinstance(lowerCamelCase , lowerCamelCase ): # Matrix multiplication assert self.column == another.row __a = Matrix(self.row , another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: __a = F"Unsupported type given for another ({type(lowerCamelCase )})" raise TypeError(lowerCamelCase ) def a__ ( self ): __a = Matrix(self.column , self.row ) for r in range(self.row ): for c in range(self.column ): __a = self[r, c] return result def a__ ( self , lowerCamelCase , lowerCamelCase ): assert isinstance(lowerCamelCase , lowerCamelCase ) and isinstance(lowerCamelCase , lowerCamelCase ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate __a = v.transpose() __a = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def _lowerCamelCase( ): # a^(-1) __a = Matrix(3 , 3 , 0 ) for i in range(3 ): __a = 1 print(F"a^(-1) is {ainv}" ) # u, v __a = Matrix(3 , 1 , 0 ) __a , __a , __a = 1, 2, -3 __a = Matrix(3 , 1 , 0 ) __a , __a , __a = 4, -2, 5 print(F"u is {u}" ) print(F"v is {v}" ) print(F"uv^T is {u * v.transpose()}" ) # Sherman Morrison print(F"(a + uv^T)^(-1) is {ainv.sherman_morrison(a , a )}" ) def _lowerCamelCase( ): import doctest doctest.testmod() testa()
261
0
import json import os import unittest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class UpperCamelCase_ ( UpperCAmelCase__ , unittest.TestCase ): '''simple docstring''' UpperCAmelCase__ = CLIPTokenizer UpperCAmelCase__ = CLIPTokenizerFast UpperCAmelCase__ = True UpperCAmelCase__ = {} UpperCAmelCase__ = False def SCREAMING_SNAKE_CASE ( self : List[str]) ->Tuple: '''simple docstring''' super().setUp() # fmt: off A__ = ['''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''lo''', '''l</w>''', '''w</w>''', '''r</w>''', '''t</w>''', '''low</w>''', '''er</w>''', '''lowest</w>''', '''newer</w>''', '''wider''', '''<unk>''', '''<|startoftext|>''', '''<|endoftext|>'''] # fmt: on A__ = dict(zip(UpperCAmelCase__ , range(len(UpperCAmelCase__)))) A__ = ['''#version: 0.2''', '''l o''', '''lo w</w>''', '''e r</w>'''] A__ = {'''unk_token''': '''<unk>'''} A__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file''']) A__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file''']) with open(self.vocab_file , '''w''' , encoding='''utf-8''') as fp: fp.write(json.dumps(UpperCAmelCase__) + '''\n''') with open(self.merges_file , '''w''' , encoding='''utf-8''') as fp: fp.write('''\n'''.join(UpperCAmelCase__)) def SCREAMING_SNAKE_CASE ( self : int , **UpperCAmelCase__ : Any) ->Any: '''simple docstring''' kwargs.update(self.special_tokens_map) return CLIPTokenizer.from_pretrained(self.tmpdirname , **UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Optional[Any] , **UpperCAmelCase__ : Union[str, Any]) ->Union[str, Any]: '''simple docstring''' kwargs.update(self.special_tokens_map) return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] , UpperCAmelCase__ : Optional[int]) ->int: '''simple docstring''' A__ = '''lower newer''' A__ = '''lower newer''' return input_text, output_text def SCREAMING_SNAKE_CASE ( self : List[Any]) ->Any: '''simple docstring''' A__ = CLIPTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map) A__ = '''lower newer''' A__ = ['''lo''', '''w''', '''er</w>''', '''n''', '''e''', '''w''', '''er</w>'''] A__ = tokenizer.tokenize(UpperCAmelCase__) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__) A__ = tokens + [tokenizer.unk_token] A__ = [10, 2, 16, 9, 3, 2, 16, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase__) , UpperCAmelCase__) @require_ftfy def SCREAMING_SNAKE_CASE ( self : Optional[int]) ->int: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})"""): A__ = self.tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__) A__ = self.rust_tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__) A__ = '''A\n\'ll 11p223RF☆ho!!to?\'d\'d\'\'d of a cat to-$\'\'d.''' A__ = tokenizer_s.tokenize(UpperCAmelCase__) A__ = tokenizer_r.tokenize(UpperCAmelCase__) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__) # Test that the tokenization is identical on an example containing a character (Latin Small Letter A # with Tilde) encoded in 2 different ways A__ = '''xa\u0303y''' + ''' ''' + '''x\xe3y''' A__ = tokenizer_s.tokenize(UpperCAmelCase__) A__ = tokenizer_r.tokenize(UpperCAmelCase__) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__) # Test that the tokenization is identical on unicode of space type A__ = [ '''\u0009''', # (horizontal tab, '\t') '''\u000B''', # (vertical tab) '''\u000C''', # (form feed) '''\u0020''', # (space, ' ') '''\u200E''', # (left-to-right mark):w '''\u200F''', # (right-to-left mark) ] for unicode_seq in spaces_unicodes: A__ = tokenizer_s.tokenize(UpperCAmelCase__) A__ = tokenizer_r.tokenize(UpperCAmelCase__) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__) # Test that the tokenization is identical on unicode of line break type A__ = [ '''\u000A''', # (line feed, '\n') '''\r\n''', # (carriage return and line feed, '\r\n') '''\u000D''', # (carriage return, '\r') '''\r''', # (carriage return, '\r') '''\u000D''', # (carriage return, '\r') '''\u2028''', # (line separator) '''\u2029''', # (paragraph separator) # "\u0085", # (next line) ] # The tokenization is not identical for the character "\u0085" (next line). The slow version using ftfy transforms # it into the Horizontal Ellipsis character "…" ("\u2026") while the fast version transforms it into a # space (and thus into an empty list). for unicode_seq in line_break_unicodes: A__ = tokenizer_s.tokenize(UpperCAmelCase__) A__ = tokenizer_r.tokenize(UpperCAmelCase__) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : List[Any]) ->int: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})"""): A__ = '''hello''' # `hello` is a token in the vocabulary of `pretrained_name` A__ = f"""{text_of_1_token} {text_of_1_token}""" A__ = self.rust_tokenizer_class.from_pretrained( UpperCAmelCase__ , use_fast=UpperCAmelCase__ , ) A__ = tokenizer_r(UpperCAmelCase__ , return_offsets_mapping=UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__) self.assertEqual(encoding.offset_mapping[0] , (0, len(UpperCAmelCase__))) self.assertEqual( encoding.offset_mapping[1] , (len(UpperCAmelCase__) + 1, len(UpperCAmelCase__) + 1 + len(UpperCAmelCase__)) , ) A__ = f""" {text}""" A__ = self.rust_tokenizer_class.from_pretrained( UpperCAmelCase__ , use_fast=UpperCAmelCase__ , ) A__ = tokenizer_r(UpperCAmelCase__ , return_offsets_mapping=UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(UpperCAmelCase__))) self.assertEqual( encoding.offset_mapping[1] , (1 + len(UpperCAmelCase__) + 1, 1 + len(UpperCAmelCase__) + 1 + len(UpperCAmelCase__)) , ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any]) ->Any: '''simple docstring''' with self.assertRaises(UpperCAmelCase__) as context: self.rust_tokenizer_class.from_pretrained('''robot-test/old-clip-tokenizer''') self.assertTrue( context.exception.args[0].startswith( '''The `backend_tokenizer` provided does not match the expected format.''')) @require_ftfy def SCREAMING_SNAKE_CASE ( self : List[str]) ->Any: '''simple docstring''' super().test_tokenization_python_rust_equals() def SCREAMING_SNAKE_CASE ( self : Optional[Any]) ->Dict: '''simple docstring''' pass
231
class UpperCamelCase_ : '''simple docstring''' def __init__( self : List[str] , UpperCAmelCase__ : int) ->Union[str, Any]: '''simple docstring''' A__ = n A__ = [None] * self.n A__ = 0 # index of the first element A__ = 0 A__ = 0 def __len__( self : List[Any]) ->int: '''simple docstring''' return self.size def SCREAMING_SNAKE_CASE ( self : List[Any]) ->bool: '''simple docstring''' return self.size == 0 def SCREAMING_SNAKE_CASE ( self : str) ->List[Any]: '''simple docstring''' return False if self.is_empty() else self.array[self.front] def SCREAMING_SNAKE_CASE ( self : List[str] , UpperCAmelCase__ : str) ->Any: '''simple docstring''' if self.size >= self.n: raise Exception('''QUEUE IS FULL''') A__ = data A__ = (self.rear + 1) % self.n self.size += 1 return self def SCREAMING_SNAKE_CASE ( self : Union[str, Any]) ->Any: '''simple docstring''' if self.size == 0: raise Exception('''UNDERFLOW''') A__ = self.array[self.front] A__ = None A__ = (self.front + 1) % self.n self.size -= 1 return temp
231
1
'''simple docstring''' def snake_case__ ( _A: str ) -> list[int]: '''simple docstring''' lowerCAmelCase = [0 for i in range(len(_A ) )] # initialize interval's left pointer and right pointer lowerCAmelCase , lowerCAmelCase = 0, 0 for i in range(1 , len(_A ) ): # case when current index is inside the interval if i <= right_pointer: lowerCAmelCase = min(right_pointer - i + 1 , z_result[i - left_pointer] ) lowerCAmelCase = min_edge while go_next(_A , _A , _A ): z_result[i] += 1 # if new index's result gives us more right interval, # we've to update left_pointer and right_pointer if i + z_result[i] - 1 > right_pointer: lowerCAmelCase , lowerCAmelCase = i, i + z_result[i] - 1 return z_result def snake_case__ ( _A: int , _A: list[int] , _A: str ) -> bool: '''simple docstring''' return i + z_result[i] < len(_A ) and s[z_result[i]] == s[i + z_result[i]] def snake_case__ ( _A: str , _A: str ) -> int: '''simple docstring''' lowerCAmelCase = 0 # concatenate 'pattern' and 'input_str' and call z_function # with concatenated string lowerCAmelCase = z_function(pattern + input_str ) for val in z_result: # if value is greater then length of the pattern string # that means this index is starting position of substring # which is equal to pattern string if val >= len(_A ): answer += 1 return answer if __name__ == "__main__": import doctest doctest.testmod()
272
'''simple docstring''' import enum import warnings from .. import MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING from ..utils import add_end_docstrings, is_tf_available from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf class a__( enum.Enum ): '''simple docstring''' UpperCAmelCase_ : Dict = 0 UpperCAmelCase_ : Dict = 1 UpperCAmelCase_ : Any = 2 @add_end_docstrings(lowerCAmelCase__ ) class a__( lowerCAmelCase__ ): '''simple docstring''' UpperCAmelCase_ : int = ''' In 1991, the remains of Russian Tsar Nicholas II and his family (except for Alexei and Maria) are discovered. The voice of Nicholas\'s young son, Tsarevich Alexei Nikolaevich, narrates the remainder of the story. 1883 Western Siberia, a young Grigori Rasputin is asked by his father and a group of men to perform magic. Rasputin has a vision and denounces one of the men as a horse thief. Although his father initially slaps him for making such an accusation, Rasputin watches as the man is chased outside and beaten. Twenty years later, Rasputin sees a vision of the Virgin Mary, prompting him to become a priest. Rasputin quickly becomes famous, with people, even a bishop, begging for his blessing. <eod> </s> <eos> ''' def __init__( self , *__lowerCAmelCase , **__lowerCAmelCase): """simple docstring""" super().__init__(*__lowerCAmelCase , **__lowerCAmelCase) self.check_model_type( TF_MODEL_FOR_CAUSAL_LM_MAPPING if self.framework == """tf""" else MODEL_FOR_CAUSAL_LM_MAPPING) if "prefix" not in self._preprocess_params: # This is very specific. The logic is quite complex and needs to be done # as a "default". # It also defines both some preprocess_kwargs and generate_kwargs # which is why we cannot put them in their respective methods. lowerCAmelCase = None if self.model.config.prefix is not None: lowerCAmelCase = self.model.config.prefix if prefix is None and self.model.__class__.__name__ in [ "XLNetLMHeadModel", "TransfoXLLMHeadModel", "TFXLNetLMHeadModel", "TFTransfoXLLMHeadModel", ]: # For XLNet and TransformerXL we add an article to the prompt to give more state to the model. lowerCAmelCase = self.XL_PREFIX if prefix is not None: # Recalculate some generate_kwargs linked to prefix. lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = self._sanitize_parameters(prefix=__lowerCAmelCase , **self._forward_params) lowerCAmelCase = {**self._preprocess_params, **preprocess_params} lowerCAmelCase = {**self._forward_params, **forward_params} def a_ ( self , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , **__lowerCAmelCase , ): """simple docstring""" lowerCAmelCase = {} if prefix is not None: lowerCAmelCase = prefix if prefix: lowerCAmelCase = self.tokenizer( __lowerCAmelCase , padding=__lowerCAmelCase , add_special_tokens=__lowerCAmelCase , return_tensors=self.framework) lowerCAmelCase = prefix_inputs["""input_ids"""].shape[-1] if handle_long_generation is not None: if handle_long_generation not in {"hole"}: raise ValueError( f"{handle_long_generation} is not a valid value for `handle_long_generation` parameter expected" """ [None, 'hole']""") lowerCAmelCase = handle_long_generation preprocess_params.update(__lowerCAmelCase) lowerCAmelCase = generate_kwargs lowerCAmelCase = {} if return_full_text is not None and return_type is None: if return_text is not None: raise ValueError("""`return_text` is mutually exclusive with `return_full_text`""") if return_tensors is not None: raise ValueError("""`return_full_text` is mutually exclusive with `return_tensors`""") lowerCAmelCase = ReturnType.FULL_TEXT if return_full_text else ReturnType.NEW_TEXT if return_tensors is not None and return_type is None: if return_text is not None: raise ValueError("""`return_text` is mutually exclusive with `return_tensors`""") lowerCAmelCase = ReturnType.TENSORS if return_type is not None: lowerCAmelCase = return_type if clean_up_tokenization_spaces is not None: lowerCAmelCase = clean_up_tokenization_spaces if stop_sequence is not None: lowerCAmelCase = self.tokenizer.encode(__lowerCAmelCase , add_special_tokens=__lowerCAmelCase) if len(__lowerCAmelCase) > 1: warnings.warn( """Stopping on a multiple token sequence is not yet supported on transformers. The first token of""" """ the stop sequence will be used as the stop sequence string in the interim.""") lowerCAmelCase = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def a_ ( self , *__lowerCAmelCase , **__lowerCAmelCase): """simple docstring""" if self.model.__class__.__name__ in ["TransfoXLLMHeadModel"]: kwargs.update({"""add_space_before_punct_symbol""": True}) return super()._parse_and_tokenize(*__lowerCAmelCase , **__lowerCAmelCase) def __call__( self , __lowerCAmelCase , **__lowerCAmelCase): """simple docstring""" return super().__call__(__lowerCAmelCase , **__lowerCAmelCase) def a_ ( self , __lowerCAmelCase , __lowerCAmelCase="" , __lowerCAmelCase=None , **__lowerCAmelCase): """simple docstring""" lowerCAmelCase = self.tokenizer( prefix + prompt_text , padding=__lowerCAmelCase , add_special_tokens=__lowerCAmelCase , return_tensors=self.framework) lowerCAmelCase = prompt_text if handle_long_generation == "hole": lowerCAmelCase = inputs["""input_ids"""].shape[-1] if "max_new_tokens" in generate_kwargs: lowerCAmelCase = generate_kwargs["""max_new_tokens"""] else: lowerCAmelCase = generate_kwargs.get("""max_length""" , self.model.config.max_length) - cur_len if new_tokens < 0: raise ValueError("""We cannot infer how many new tokens are expected""") if cur_len + new_tokens > self.tokenizer.model_max_length: lowerCAmelCase = self.tokenizer.model_max_length - new_tokens if keep_length <= 0: raise ValueError( """We cannot use `hole` to handle this generation the number of desired tokens exceeds the""" """ models max length""") lowerCAmelCase = inputs["""input_ids"""][:, -keep_length:] if "attention_mask" in inputs: lowerCAmelCase = inputs["""attention_mask"""][:, -keep_length:] return inputs def a_ ( self , __lowerCAmelCase , **__lowerCAmelCase): """simple docstring""" lowerCAmelCase = model_inputs["""input_ids"""] lowerCAmelCase = model_inputs.get("""attention_mask""" , __lowerCAmelCase) # Allow empty prompts if input_ids.shape[1] == 0: lowerCAmelCase = None lowerCAmelCase = None lowerCAmelCase = 1 else: lowerCAmelCase = input_ids.shape[0] lowerCAmelCase = model_inputs.pop("""prompt_text""") # If there is a prefix, we may need to adjust the generation length. Do so without permanently modifying # generate_kwargs, as some of the parameterization may come from the initialization of the pipeline. lowerCAmelCase = generate_kwargs.pop("""prefix_length""" , 0) if prefix_length > 0: lowerCAmelCase = """max_new_tokens""" in generate_kwargs or ( """generation_config""" in generate_kwargs and generate_kwargs["""generation_config"""].max_new_tokens is not None ) if not has_max_new_tokens: lowerCAmelCase = generate_kwargs.get("""max_length""") or self.model.config.max_length generate_kwargs["max_length"] += prefix_length lowerCAmelCase = """min_new_tokens""" in generate_kwargs or ( """generation_config""" in generate_kwargs and generate_kwargs["""generation_config"""].min_new_tokens is not None ) if not has_min_new_tokens and "min_length" in generate_kwargs: generate_kwargs["min_length"] += prefix_length # BS x SL lowerCAmelCase = self.model.generate(input_ids=__lowerCAmelCase , attention_mask=__lowerCAmelCase , **__lowerCAmelCase) lowerCAmelCase = generated_sequence.shape[0] if self.framework == "pt": lowerCAmelCase = generated_sequence.reshape(__lowerCAmelCase , out_b // in_b , *generated_sequence.shape[1:]) elif self.framework == "tf": lowerCAmelCase = tf.reshape(__lowerCAmelCase , (in_b, out_b // in_b, *generated_sequence.shape[1:])) return {"generated_sequence": generated_sequence, "input_ids": input_ids, "prompt_text": prompt_text} def a_ ( self , __lowerCAmelCase , __lowerCAmelCase=ReturnType.FULL_TEXT , __lowerCAmelCase=True): """simple docstring""" lowerCAmelCase = model_outputs["""generated_sequence"""][0] lowerCAmelCase = model_outputs["""input_ids"""] lowerCAmelCase = model_outputs["""prompt_text"""] lowerCAmelCase = generated_sequence.numpy().tolist() lowerCAmelCase = [] for sequence in generated_sequence: if return_type == ReturnType.TENSORS: lowerCAmelCase = {"""generated_token_ids""": sequence} elif return_type in {ReturnType.NEW_TEXT, ReturnType.FULL_TEXT}: # Decode text lowerCAmelCase = self.tokenizer.decode( __lowerCAmelCase , skip_special_tokens=__lowerCAmelCase , clean_up_tokenization_spaces=__lowerCAmelCase , ) # Remove PADDING prompt of the sequence if XLNet or Transfo-XL model is used if input_ids is None: lowerCAmelCase = 0 else: lowerCAmelCase = len( self.tokenizer.decode( input_ids[0] , skip_special_tokens=__lowerCAmelCase , clean_up_tokenization_spaces=__lowerCAmelCase , )) if return_type == ReturnType.FULL_TEXT: lowerCAmelCase = prompt_text + text[prompt_length:] else: lowerCAmelCase = text[prompt_length:] lowerCAmelCase = {"""generated_text""": all_text} records.append(__lowerCAmelCase) return records
272
1
'''simple docstring''' __a: Dict = { """A""": """.-""", """B""": """-...""", """C""": """-.-.""", """D""": """-..""", """E""": """.""", """F""": """..-.""", """G""": """--.""", """H""": """....""", """I""": """..""", """J""": """.---""", """K""": """-.-""", """L""": """.-..""", """M""": """--""", """N""": """-.""", """O""": """---""", """P""": """.--.""", """Q""": """--.-""", """R""": """.-.""", """S""": """...""", """T""": """-""", """U""": """..-""", """V""": """...-""", """W""": """.--""", """X""": """-..-""", """Y""": """-.--""", """Z""": """--..""", """1""": """.----""", """2""": """..---""", """3""": """...--""", """4""": """....-""", """5""": """.....""", """6""": """-....""", """7""": """--...""", """8""": """---..""", """9""": """----.""", """0""": """-----""", """&""": """.-...""", """@""": """.--.-.""", """:""": """---...""", """,""": """--..--""", """.""": """.-.-.-""", """'""": """.----.""", """\"""": """.-..-.""", """?""": """..--..""", """/""": """-..-.""", """=""": """-...-""", """+""": """.-.-.""", """-""": """-....-""", """(""": """-.--.""", """)""": """-.--.-""", """!""": """-.-.--""", """ """: """/""" } # Exclamation mark is not in ITU-R recommendation # fmt: on __a: str = {value: key for key, value in MORSE_CODE_DICT.items()} def __UpperCamelCase ( UpperCAmelCase ): return " ".join(MORSE_CODE_DICT[char] for char in message.upper() ) def __UpperCamelCase ( UpperCAmelCase ): return "".join(REVERSE_DICT[char] for char in message.split() ) def __UpperCamelCase ( ): lowercase__ : Tuple = '''Morse code here!''' print(UpperCAmelCase ) lowercase__ : int = encrypt(UpperCAmelCase ) print(UpperCAmelCase ) lowercase__ : Any = decrypt(UpperCAmelCase ) print(UpperCAmelCase ) if __name__ == "__main__": main()
214
'''simple docstring''' class UpperCAmelCase : '''simple docstring''' def __init__( self ) -> List[str]: lowercase__ : Dict = {} def _lowerCAmelCase( self ) -> None: print(self.vertex ) for i in self.vertex: print(__lowerCAmelCase , ''' -> ''' , ''' -> '''.join([str(__lowerCAmelCase ) for j in self.vertex[i]] ) ) def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase ) -> None: # check if vertex is already present, if from_vertex in self.vertex: self.vertex[from_vertex].append(__lowerCAmelCase ) else: # else make a new vertex lowercase__ : Union[str, Any] = [to_vertex] def _lowerCAmelCase( self ) -> None: # visited array for storing already visited nodes lowercase__ : str = [False] * len(self.vertex ) # call the recursive helper function for i in range(len(self.vertex ) ): if not visited[i]: self.dfs_recursive(__lowerCAmelCase , __lowerCAmelCase ) def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase ) -> None: # mark start vertex as visited lowercase__ : List[str] = True print(__lowerCAmelCase , end=''' ''' ) # Recur for all the vertices that are adjacent to this node for i in self.vertex: if not visited[i]: self.dfs_recursive(__lowerCAmelCase , __lowerCAmelCase ) if __name__ == "__main__": __a: Optional[Any] = Graph() g.add_edge(0, 1) g.add_edge(0, 2) g.add_edge(1, 2) g.add_edge(2, 0) g.add_edge(2, 3) g.add_edge(3, 3) g.print_graph() print("""DFS:""") g.dfs() # OUTPUT: # 0 -> 1 -> 2 # 1 -> 2 # 2 -> 0 -> 3 # 3 -> 3 # DFS: # 0 1 2 3
214
1
from __future__ import annotations _SCREAMING_SNAKE_CASE : Optional[int] = [] def UpperCAmelCase__ (UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_ ): """simple docstring""" for i in range(len(UpperCamelCase_ ) ): if board[row][i] == 1: return False for i in range(len(UpperCamelCase_ ) ): if board[i][column] == 1: return False for i, j in zip(range(UpperCamelCase_ ,-1 ,-1 ) ,range(UpperCamelCase_ ,-1 ,-1 ) ): if board[i][j] == 1: return False for i, j in zip(range(UpperCamelCase_ ,-1 ,-1 ) ,range(UpperCamelCase_ ,len(UpperCamelCase_ ) ) ): if board[i][j] == 1: return False return True def UpperCAmelCase__ (UpperCamelCase_ ,UpperCamelCase_ ): """simple docstring""" if row >= len(UpperCamelCase_ ): solution.append(UpperCamelCase_ ) printboard(UpperCamelCase_ ) print() return True for i in range(len(UpperCamelCase_ ) ): if is_safe(UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_ ): snake_case = 1 solve(UpperCamelCase_ ,row + 1 ) snake_case = 0 return False def UpperCAmelCase__ (UpperCamelCase_ ): """simple docstring""" for i in range(len(UpperCamelCase_ ) ): for j in range(len(UpperCamelCase_ ) ): if board[i][j] == 1: print('''Q''' ,end=''' ''' ) else: print('''.''' ,end=''' ''' ) print() # n=int(input("The no. of queens")) _SCREAMING_SNAKE_CASE : Tuple = 8 _SCREAMING_SNAKE_CASE : List[Any] = [[0 for i in range(n)] for j in range(n)] solve(board, 0) print("The total no. of solutions are :", len(solution))
127
import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEIGHTS_NAME, cached_file, get_file_from_repo, has_file, ) _SCREAMING_SNAKE_CASE : List[str] = "hf-internal-testing/tiny-random-bert" _SCREAMING_SNAKE_CASE : Optional[Any] = os.path.join(TRANSFORMERS_CACHE, "models--hf-internal-testing--tiny-random-bert") _SCREAMING_SNAKE_CASE : Optional[int] = "9b8c223d42b2188cb49d29af482996f9d0f3e5a6" class A__ ( unittest.TestCase ): """simple docstring""" def a_ ( self ): snake_case = cached_file(__snake_case , __snake_case ) # Should have downloaded the file in here self.assertTrue(os.path.isdir(__snake_case ) ) # Cache should contain at least those three subfolders: for subfolder in ["blobs", "refs", "snapshots"]: self.assertTrue(os.path.isdir(os.path.join(__snake_case , __snake_case ) ) ) with open(os.path.join(__snake_case , '''refs''' , '''main''' ) ) as f: snake_case = f.read() self.assertEqual(__snake_case , os.path.join(__snake_case , '''snapshots''' , __snake_case , __snake_case ) ) self.assertTrue(os.path.isfile(__snake_case ) ) # File is cached at the same place the second time. snake_case = cached_file(__snake_case , __snake_case ) self.assertEqual(__snake_case , __snake_case ) # Using a specific revision to test the full commit hash. snake_case = cached_file(__snake_case , __snake_case , revision='''9b8c223''' ) self.assertEqual(__snake_case , os.path.join(__snake_case , '''snapshots''' , __snake_case , __snake_case ) ) def a_ ( self ): with self.assertRaisesRegex(__snake_case , '''is not a valid model identifier''' ): snake_case = cached_file('''tiny-random-bert''' , __snake_case ) with self.assertRaisesRegex(__snake_case , '''is not a valid git identifier''' ): snake_case = cached_file(__snake_case , __snake_case , revision='''aaaa''' ) with self.assertRaisesRegex(__snake_case , '''does not appear to have a file named''' ): snake_case = cached_file(__snake_case , '''conf''' ) def a_ ( self ): with self.assertRaisesRegex(__snake_case , '''does not appear to have a file named''' ): snake_case = cached_file(__snake_case , '''conf''' ) with open(os.path.join(__snake_case , '''refs''' , '''main''' ) ) as f: snake_case = f.read() self.assertTrue(os.path.isfile(os.path.join(__snake_case , '''.no_exist''' , __snake_case , '''conf''' ) ) ) snake_case = cached_file(__snake_case , '''conf''' , _raise_exceptions_for_missing_entries=__snake_case ) self.assertIsNone(__snake_case ) snake_case = cached_file(__snake_case , '''conf''' , local_files_only=__snake_case , _raise_exceptions_for_missing_entries=__snake_case ) self.assertIsNone(__snake_case ) snake_case = mock.Mock() snake_case = 5_0_0 snake_case = {} snake_case = HTTPError snake_case = {} # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('''requests.Session.request''' , return_value=__snake_case ) as mock_head: snake_case = cached_file(__snake_case , '''conf''' , _raise_exceptions_for_connection_errors=__snake_case ) self.assertIsNone(__snake_case ) # This check we did call the fake head request mock_head.assert_called() def a_ ( self ): self.assertTrue(has_file('''hf-internal-testing/tiny-bert-pt-only''' , __snake_case ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , __snake_case ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , __snake_case ) ) def a_ ( self ): # `get_file_from_repo` returns None if the file does not exist self.assertIsNone(get_file_from_repo('''bert-base-cased''' , '''ahah.txt''' ) ) # The function raises if the repository does not exist. with self.assertRaisesRegex(__snake_case , '''is not a valid model identifier''' ): get_file_from_repo('''bert-base-case''' , __snake_case ) # The function raises if the revision does not exist. with self.assertRaisesRegex(__snake_case , '''is not a valid git identifier''' ): get_file_from_repo('''bert-base-cased''' , __snake_case , revision='''ahaha''' ) snake_case = get_file_from_repo('''bert-base-cased''' , __snake_case ) # The name is the cached name which is not very easy to test, so instead we load the content. snake_case = json.loads(open(__snake_case , '''r''' ).read() ) self.assertEqual(config['''hidden_size'''] , 7_6_8 ) def a_ ( self ): with tempfile.TemporaryDirectory() as tmp_dir: snake_case = Path(__snake_case ) / '''a.txt''' filename.touch() self.assertEqual(get_file_from_repo(__snake_case , '''a.txt''' ) , str(__snake_case ) ) self.assertIsNone(get_file_from_repo(__snake_case , '''b.txt''' ) )
127
1
"""simple docstring""" import os import unittest from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, BertTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class UpperCAmelCase_ ( A_, unittest.TestCase ): lowercase__ = BertTokenizer lowercase__ = BertTokenizerFast lowercase__ = True lowercase__ = True lowercase__ = filter_non_english def __magic_name__ ( self : List[Any] ) -> Dict: '''simple docstring''' super().setUp() A__ = [ "[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] A__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def __magic_name__ ( self : Any , snake_case_ : List[str] ) -> Union[str, Any]: '''simple docstring''' A__ = "UNwant\u00E9d,running" A__ = "unwanted, running" return input_text, output_text def __magic_name__ ( self : str ) -> Optional[int]: '''simple docstring''' A__ = self.tokenizer_class(self.vocab_file ) A__ = tokenizer.tokenize("UNwant\u00E9d,running" ) self.assertListEqual(snake_case_ , ["un", "##want", "##ed", ",", "runn", "##ing"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(snake_case_ ) , [9, 6, 7, 12, 10, 11] ) def __magic_name__ ( self : Optional[Any] ) -> str: '''simple docstring''' if not self.test_rust_tokenizer: return A__ = self.get_tokenizer() A__ = self.get_rust_tokenizer() A__ = "UNwant\u00E9d,running" A__ = tokenizer.tokenize(snake_case_ ) A__ = rust_tokenizer.tokenize(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) A__ = tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) A__ = rust_tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) A__ = self.get_rust_tokenizer() A__ = tokenizer.encode(snake_case_ ) A__ = rust_tokenizer.encode(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) # With lower casing A__ = self.get_tokenizer(do_lower_case=snake_case_ ) A__ = self.get_rust_tokenizer(do_lower_case=snake_case_ ) A__ = "UNwant\u00E9d,running" A__ = tokenizer.tokenize(snake_case_ ) A__ = rust_tokenizer.tokenize(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) A__ = tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) A__ = rust_tokenizer.encode(snake_case_ , add_special_tokens=snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) A__ = self.get_rust_tokenizer() A__ = tokenizer.encode(snake_case_ ) A__ = rust_tokenizer.encode(snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) def __magic_name__ ( self : List[str] ) -> Optional[Any]: '''simple docstring''' A__ = BasicTokenizer() self.assertListEqual(tokenizer.tokenize("ah\u535A\u63A8zz" ) , ["ah", "\u535A", "\u63A8", "zz"] ) def __magic_name__ ( self : int ) -> str: '''simple docstring''' A__ = BasicTokenizer(do_lower_case=snake_case_ ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? " ) , ["hello", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["hello"] ) def __magic_name__ ( self : Dict ) -> List[Any]: '''simple docstring''' A__ = BasicTokenizer(do_lower_case=snake_case_ , strip_accents=snake_case_ ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["hällo", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["h\u00E9llo"] ) def __magic_name__ ( self : Optional[int] ) -> int: '''simple docstring''' A__ = BasicTokenizer(do_lower_case=snake_case_ , strip_accents=snake_case_ ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["hallo", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["hello"] ) def __magic_name__ ( self : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' A__ = BasicTokenizer(do_lower_case=snake_case_ ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["hallo", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["hello"] ) def __magic_name__ ( self : List[Any] ) -> List[str]: '''simple docstring''' A__ = BasicTokenizer(do_lower_case=snake_case_ ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? " ) , ["HeLLo", "!", "how", "Are", "yoU", "?"] ) def __magic_name__ ( self : Union[str, Any] ) -> Dict: '''simple docstring''' A__ = BasicTokenizer(do_lower_case=snake_case_ , strip_accents=snake_case_ ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["HäLLo", "!", "how", "Are", "yoU", "?"] ) def __magic_name__ ( self : List[Any] ) -> Optional[int]: '''simple docstring''' A__ = BasicTokenizer(do_lower_case=snake_case_ , strip_accents=snake_case_ ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["HaLLo", "!", "how", "Are", "yoU", "?"] ) def __magic_name__ ( self : Optional[int] ) -> Dict: '''simple docstring''' A__ = BasicTokenizer(do_lower_case=snake_case_ , never_split=["[UNK]"] ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? [UNK]" ) , ["HeLLo", "!", "how", "Are", "yoU", "?", "[UNK]"] ) def __magic_name__ ( self : int ) -> Optional[Any]: '''simple docstring''' A__ = BasicTokenizer() A__ = "a\n'll !!to?'d of, can't." A__ = ["a", "'", "ll", "!", "!", "to", "?", "'", "d", "of", ",", "can", "'", "t", "."] self.assertListEqual(tokenizer.tokenize(snake_case_ ) , snake_case_ ) def __magic_name__ ( self : Dict ) -> Union[str, Any]: '''simple docstring''' A__ = ["[UNK]", "[CLS]", "[SEP]", "want", "##want", "##ed", "wa", "un", "runn", "##ing"] A__ = {} for i, token in enumerate(snake_case_ ): A__ = i A__ = WordpieceTokenizer(vocab=snake_case_ , unk_token="[UNK]" ) self.assertListEqual(tokenizer.tokenize("" ) , [] ) self.assertListEqual(tokenizer.tokenize("unwanted running" ) , ["un", "##want", "##ed", "runn", "##ing"] ) self.assertListEqual(tokenizer.tokenize("unwantedX running" ) , ["[UNK]", "runn", "##ing"] ) def __magic_name__ ( self : List[Any] ) -> Tuple: '''simple docstring''' self.assertTrue(_is_whitespace(" " ) ) self.assertTrue(_is_whitespace("\t" ) ) self.assertTrue(_is_whitespace("\r" ) ) self.assertTrue(_is_whitespace("\n" ) ) self.assertTrue(_is_whitespace("\u00A0" ) ) self.assertFalse(_is_whitespace("A" ) ) self.assertFalse(_is_whitespace("-" ) ) def __magic_name__ ( self : Tuple ) -> str: '''simple docstring''' self.assertTrue(_is_control("\u0005" ) ) self.assertFalse(_is_control("A" ) ) self.assertFalse(_is_control(" " ) ) self.assertFalse(_is_control("\t" ) ) self.assertFalse(_is_control("\r" ) ) def __magic_name__ ( self : List[str] ) -> Optional[int]: '''simple docstring''' self.assertTrue(_is_punctuation("-" ) ) self.assertTrue(_is_punctuation("$" ) ) self.assertTrue(_is_punctuation("`" ) ) self.assertTrue(_is_punctuation("." ) ) self.assertFalse(_is_punctuation("A" ) ) self.assertFalse(_is_punctuation(" " ) ) def __magic_name__ ( self : str ) -> Tuple: '''simple docstring''' A__ = self.get_tokenizer() A__ = self.get_rust_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(snake_case_ ) for t in ["Test", "\xad", "test"]] , [["[UNK]"], [], ["[UNK]"]] ) self.assertListEqual( [rust_tokenizer.tokenize(snake_case_ ) for t in ["Test", "\xad", "test"]] , [["[UNK]"], [], ["[UNK]"]] ) @slow def __magic_name__ ( self : Any ) -> int: '''simple docstring''' A__ = self.tokenizer_class.from_pretrained("bert-base-uncased" ) A__ = tokenizer.encode("sequence builders" , add_special_tokens=snake_case_ ) A__ = tokenizer.encode("multi-sequence build" , add_special_tokens=snake_case_ ) A__ = tokenizer.build_inputs_with_special_tokens(snake_case_ ) A__ = tokenizer.build_inputs_with_special_tokens(snake_case_ , snake_case_ ) assert encoded_sentence == [101] + text + [102] assert encoded_pair == [101] + text + [102] + text_a + [102] def __magic_name__ ( self : str ) -> str: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): A__ = self.rust_tokenizer_class.from_pretrained(snake_case_ , **snake_case_ ) A__ = F"""A, naïve {tokenizer_r.mask_token} AllenNLP sentence.""" A__ = tokenizer_r.encode_plus( snake_case_ , return_attention_mask=snake_case_ , return_token_type_ids=snake_case_ , return_offsets_mapping=snake_case_ , add_special_tokens=snake_case_ , ) A__ = tokenizer_r.do_lower_case if hasattr(snake_case_ , "do_lower_case" ) else False A__ = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), "A"), ((1, 2), ","), ((3, 5), "na"), ((5, 6), "##ï"), ((6, 8), "##ve"), ((9, 15), tokenizer_r.mask_token), ((16, 21), "Allen"), ((21, 23), "##NL"), ((23, 24), "##P"), ((25, 33), "sentence"), ((33, 34), "."), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), "a"), ((1, 2), ","), ((3, 8), "naive"), ((9, 15), tokenizer_r.mask_token), ((16, 21), "allen"), ((21, 23), "##nl"), ((23, 24), "##p"), ((25, 33), "sentence"), ((33, 34), "."), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens["input_ids"] ) ) self.assertEqual([e[0] for e in expected_results] , tokens["offset_mapping"] ) def __magic_name__ ( self : Optional[Any] ) -> int: '''simple docstring''' A__ = ["的", "人", "有"] A__ = "".join(snake_case_ ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): A__ = True A__ = self.tokenizer_class.from_pretrained(snake_case_ , **snake_case_ ) A__ = self.rust_tokenizer_class.from_pretrained(snake_case_ , **snake_case_ ) A__ = tokenizer_p.encode(snake_case_ , add_special_tokens=snake_case_ ) A__ = tokenizer_r.encode(snake_case_ , add_special_tokens=snake_case_ ) A__ = tokenizer_r.convert_ids_to_tokens(snake_case_ ) A__ = tokenizer_p.convert_ids_to_tokens(snake_case_ ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(snake_case_ , snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ ) A__ = False A__ = self.rust_tokenizer_class.from_pretrained(snake_case_ , **snake_case_ ) A__ = self.tokenizer_class.from_pretrained(snake_case_ , **snake_case_ ) A__ = tokenizer_r.encode(snake_case_ , add_special_tokens=snake_case_ ) A__ = tokenizer_p.encode(snake_case_ , add_special_tokens=snake_case_ ) A__ = tokenizer_r.convert_ids_to_tokens(snake_case_ ) A__ = tokenizer_p.convert_ids_to_tokens(snake_case_ ) # it is expected that only the first Chinese character is not preceded by "##". A__ = [ F"""##{token}""" if idx != 0 else token for idx, token in enumerate(snake_case_ ) ] self.assertListEqual(snake_case_ , snake_case_ ) self.assertListEqual(snake_case_ , snake_case_ )
230
"""simple docstring""" from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class UpperCAmelCase_ ( A_ ): lowercase__ = ['''image_processor''', '''tokenizer'''] lowercase__ = '''AutoImageProcessor''' lowercase__ = '''AutoTokenizer''' def __init__( self : str , snake_case_ : Dict , snake_case_ : List[str] ) -> str: '''simple docstring''' super().__init__(snake_case_ , snake_case_ ) A__ = self.image_processor def __call__( self : int , snake_case_ : Any=None , snake_case_ : Any=None , snake_case_ : Union[str, Any]=None , **snake_case_ : Optional[int] ) -> Optional[int]: '''simple docstring''' if text is None and images is None: raise ValueError("You have to specify either text or images. Both cannot be none." ) if text is not None: A__ = self.tokenizer(snake_case_ , return_tensors=snake_case_ , **snake_case_ ) if images is not None: A__ = self.image_processor(snake_case_ , return_tensors=snake_case_ , **snake_case_ ) if text is not None and images is not None: A__ = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**snake_case_ ) , tensor_type=snake_case_ ) def __magic_name__ ( self : Optional[int] , *snake_case_ : Union[str, Any] , **snake_case_ : List[Any] ) -> int: '''simple docstring''' return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def __magic_name__ ( self : List[str] , *snake_case_ : List[str] , **snake_case_ : Optional[int] ) -> Tuple: '''simple docstring''' return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property def __magic_name__ ( self : List[Any] ) -> List[Any]: '''simple docstring''' return ["input_ids", "attention_mask", "pixel_values"]
230
1
import copy from collections import OrderedDict from typing import Dict, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _UpperCAmelCase : Optional[int] = logging.get_logger(__name__) _UpperCAmelCase : Any = { "facebook/detr-resnet-50": "https://huggingface.co/facebook/detr-resnet-50/resolve/main/config.json", # See all DETR models at https://huggingface.co/models?filter=detr } class lowercase ( _lowerCAmelCase ): __SCREAMING_SNAKE_CASE : Union[str, Any] = '''detr''' __SCREAMING_SNAKE_CASE : Union[str, Any] = ['''past_key_values'''] __SCREAMING_SNAKE_CASE : Dict = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', } def __init__( self , snake_case=True , snake_case=None , snake_case=3 , snake_case=100 , snake_case=6 , snake_case=2048 , snake_case=8 , snake_case=6 , snake_case=2048 , snake_case=8 , snake_case=0.0 , snake_case=0.0 , snake_case=True , snake_case="relu" , snake_case=256 , snake_case=0.1 , snake_case=0.0 , snake_case=0.0 , snake_case=0.02 , snake_case=1.0 , snake_case=False , snake_case="sine" , snake_case="resnet50" , snake_case=True , snake_case=False , snake_case=1 , snake_case=5 , snake_case=2 , snake_case=1 , snake_case=1 , snake_case=5 , snake_case=2 , snake_case=0.1 , **snake_case , ): if backbone_config is not None and use_timm_backbone: raise ValueError('You can\'t specify both `backbone_config` and `use_timm_backbone`.' ) if not use_timm_backbone: if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.' ) snake_case_ = CONFIG_MAPPING["""resnet"""](out_features=['stage4'] ) elif isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): snake_case_ = backbone_config.get('model_type' ) snake_case_ = CONFIG_MAPPING[backbone_model_type] snake_case_ = config_class.from_dict(lowerCAmelCase__ ) # set timm attributes to None snake_case_ = None, None, None snake_case_ = use_timm_backbone snake_case_ = backbone_config snake_case_ = num_channels snake_case_ = num_queries snake_case_ = d_model snake_case_ = encoder_ffn_dim snake_case_ = encoder_layers snake_case_ = encoder_attention_heads snake_case_ = decoder_ffn_dim snake_case_ = decoder_layers snake_case_ = decoder_attention_heads snake_case_ = dropout snake_case_ = attention_dropout snake_case_ = activation_dropout snake_case_ = activation_function snake_case_ = init_std snake_case_ = init_xavier_std snake_case_ = encoder_layerdrop snake_case_ = decoder_layerdrop snake_case_ = encoder_layers snake_case_ = auxiliary_loss snake_case_ = position_embedding_type snake_case_ = backbone snake_case_ = use_pretrained_backbone snake_case_ = dilation # Hungarian matcher snake_case_ = class_cost snake_case_ = bbox_cost snake_case_ = giou_cost # Loss coefficients snake_case_ = mask_loss_coefficient snake_case_ = dice_loss_coefficient snake_case_ = bbox_loss_coefficient snake_case_ = giou_loss_coefficient snake_case_ = eos_coefficient super().__init__(is_encoder_decoder=lowerCAmelCase__ , **lowerCAmelCase__ ) @property def a ( self ): return self.encoder_attention_heads @property def a ( self ): return self.d_model @classmethod def a ( cls , snake_case , **snake_case ): return cls(backbone_config=lowerCAmelCase__ , **lowerCAmelCase__ ) def a ( self ): snake_case_ = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: snake_case_ = self.backbone_config.to_dict() snake_case_ = self.__class__.model_type return output class lowercase ( _lowerCAmelCase ): __SCREAMING_SNAKE_CASE : Union[str, Any] = version.parse('''1.11''' ) @property def a ( self ): return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ('pixel_mask', {0: 'batch'}), ] ) @property def a ( self ): return 1e-5 @property def a ( self ): return 12
285
from __future__ import annotations def UpperCamelCase ( _A ): # This function is recursive """simple docstring""" __magic_name__ : str = len(_A ) # If the array contains only one element, we return it (it's the stop condition of # recursion) if array_length <= 1: return array # Else __magic_name__ : Dict = array[0] __magic_name__ : Optional[Any] = False __magic_name__ : Tuple = 1 __magic_name__ : list[int] = [] while not is_found and i < array_length: if array[i] < pivot: __magic_name__ : Union[str, Any] = True __magic_name__ : List[Any] = [element for element in array[i:] if element >= array[i]] __magic_name__ : Dict = longest_subsequence(_A ) if len(_A ) > len(_A ): __magic_name__ : Tuple = temp_array else: i += 1 __magic_name__ : Any = [element for element in array[1:] if element >= pivot] __magic_name__ : Dict = [pivot, *longest_subsequence(_A )] if len(_A ) > len(_A ): return temp_array else: return longest_subseq if __name__ == "__main__": import doctest doctest.testmod()
342
0
from typing import Optional, Tuple, Union import torch from einops import rearrange, reduce from diffusers import DDIMScheduler, DDPMScheduler, DiffusionPipeline, ImagePipelineOutput, UNetaDConditionModel from diffusers.schedulers.scheduling_ddim import DDIMSchedulerOutput from diffusers.schedulers.scheduling_ddpm import DDPMSchedulerOutput __UpperCamelCase : Dict = 8 def _a ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : Dict=BITS ): """simple docstring""" UpperCamelCase__ : Any = x.device UpperCamelCase__ : Optional[int] = (x * 255).int().clamp(0 , 255 ) UpperCamelCase__ : int = 2 ** torch.arange(bits - 1 , -1 , -1 , device=SCREAMING_SNAKE_CASE ) UpperCamelCase__ : str = rearrange(SCREAMING_SNAKE_CASE , '''d -> d 1 1''' ) UpperCamelCase__ : List[Any] = rearrange(SCREAMING_SNAKE_CASE , '''b c h w -> b c 1 h w''' ) UpperCamelCase__ : List[str] = ((x & mask) != 0).float() UpperCamelCase__ : Dict = rearrange(SCREAMING_SNAKE_CASE , '''b c d h w -> b (c d) h w''' ) UpperCamelCase__ : Optional[int] = bits * 2 - 1 return bits def _a ( SCREAMING_SNAKE_CASE : Dict , SCREAMING_SNAKE_CASE : Any=BITS ): """simple docstring""" UpperCamelCase__ : List[Any] = x.device UpperCamelCase__ : Union[str, Any] = (x > 0).int() UpperCamelCase__ : Tuple = 2 ** torch.arange(bits - 1 , -1 , -1 , device=SCREAMING_SNAKE_CASE , dtype=torch.intaa ) UpperCamelCase__ : Dict = rearrange(SCREAMING_SNAKE_CASE , '''d -> d 1 1''' ) UpperCamelCase__ : Dict = rearrange(SCREAMING_SNAKE_CASE , '''b (c d) h w -> b c d h w''' , d=8 ) UpperCamelCase__ : Any = reduce(x * mask , '''b c d h w -> b c h w''' , '''sum''' ) return (dec / 255).clamp(0.0 , 1.0 ) def _a ( self : Optional[int] , SCREAMING_SNAKE_CASE : torch.FloatTensor , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : torch.FloatTensor , SCREAMING_SNAKE_CASE : float = 0.0 , SCREAMING_SNAKE_CASE : bool = True , SCREAMING_SNAKE_CASE : List[Any]=None , SCREAMING_SNAKE_CASE : bool = True , ): """simple docstring""" if self.num_inference_steps is None: raise ValueError( '''Number of inference steps is \'None\', you need to run \'set_timesteps\' after creating the scheduler''' ) # See formulas (12) and (16) of DDIM paper https://arxiv.org/pdf/2010.02502.pdf # Ideally, read DDIM paper in-detail understanding # Notation (<variable name> -> <name in paper> # - pred_noise_t -> e_theta(x_t, t) # - pred_original_sample -> f_theta(x_t, t) or x_0 # - std_dev_t -> sigma_t # - eta -> η # - pred_sample_direction -> "direction pointing to x_t" # - pred_prev_sample -> "x_t-1" # 1. get previous step value (=t-1) UpperCamelCase__ : List[Any] = timestep - self.config.num_train_timesteps // self.num_inference_steps # 2. compute alphas, betas UpperCamelCase__ : Tuple = self.alphas_cumprod[timestep] UpperCamelCase__ : Dict = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.final_alpha_cumprod UpperCamelCase__ : Dict = 1 - alpha_prod_t # 3. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf UpperCamelCase__ : Optional[int] = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 # 4. Clip "predicted x_0" UpperCamelCase__ : Optional[int] = self.bit_scale if self.config.clip_sample: UpperCamelCase__ : str = torch.clamp(SCREAMING_SNAKE_CASE , -scale , SCREAMING_SNAKE_CASE ) # 5. compute variance: "sigma_t(η)" -> see formula (16) # σ_t = sqrt((1 − α_t−1)/(1 − α_t)) * sqrt(1 − α_t/α_t−1) UpperCamelCase__ : List[str] = self._get_variance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) UpperCamelCase__ : int = eta * variance ** 0.5 if use_clipped_model_output: # the model_output is always re-derived from the clipped x_0 in Glide UpperCamelCase__ : Optional[Any] = (sample - alpha_prod_t ** 0.5 * pred_original_sample) / beta_prod_t ** 0.5 # 6. compute "direction pointing to x_t" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf UpperCamelCase__ : Tuple = (1 - alpha_prod_t_prev - std_dev_t**2) ** 0.5 * model_output # 7. compute x_t without "random noise" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf UpperCamelCase__ : Union[str, Any] = alpha_prod_t_prev ** 0.5 * pred_original_sample + pred_sample_direction if eta > 0: # randn_like does not support generator https://github.com/pytorch/pytorch/issues/27072 UpperCamelCase__ : Optional[int] = model_output.device if torch.is_tensor(SCREAMING_SNAKE_CASE ) else '''cpu''' UpperCamelCase__ : Any = torch.randn(model_output.shape , dtype=model_output.dtype , generator=SCREAMING_SNAKE_CASE ).to(SCREAMING_SNAKE_CASE ) UpperCamelCase__ : Any = self._get_variance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ** 0.5 * eta * noise UpperCamelCase__ : str = prev_sample + variance if not return_dict: return (prev_sample,) return DDIMSchedulerOutput(prev_sample=SCREAMING_SNAKE_CASE , pred_original_sample=SCREAMING_SNAKE_CASE ) def _a ( self : Union[str, Any] , SCREAMING_SNAKE_CASE : torch.FloatTensor , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : torch.FloatTensor , SCREAMING_SNAKE_CASE : Dict="epsilon" , SCREAMING_SNAKE_CASE : List[str]=None , SCREAMING_SNAKE_CASE : bool = True , ): """simple docstring""" UpperCamelCase__ : Optional[int] = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type in ["learned", "learned_range"]: UpperCamelCase__ , UpperCamelCase__ : Optional[int] = torch.split(SCREAMING_SNAKE_CASE , sample.shape[1] , dim=1 ) else: UpperCamelCase__ : Optional[Any] = None # 1. compute alphas, betas UpperCamelCase__ : Tuple = self.alphas_cumprod[t] UpperCamelCase__ : Any = self.alphas_cumprod[t - 1] if t > 0 else self.one UpperCamelCase__ : Optional[int] = 1 - alpha_prod_t UpperCamelCase__ : Optional[int] = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if prediction_type == "epsilon": UpperCamelCase__ : Dict = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif prediction_type == "sample": UpperCamelCase__ : Optional[Any] = model_output else: raise ValueError(F"Unsupported prediction_type {prediction_type}." ) # 3. Clip "predicted x_0" UpperCamelCase__ : Union[str, Any] = self.bit_scale if self.config.clip_sample: UpperCamelCase__ : Tuple = torch.clamp(SCREAMING_SNAKE_CASE , -scale , SCREAMING_SNAKE_CASE ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCamelCase__ : List[Any] = (alpha_prod_t_prev ** 0.5 * self.betas[t]) / beta_prod_t UpperCamelCase__ : str = self.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCamelCase__ : int = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise UpperCamelCase__ : Optional[int] = 0 if t > 0: UpperCamelCase__ : Dict = torch.randn( model_output.size() , dtype=model_output.dtype , layout=model_output.layout , generator=SCREAMING_SNAKE_CASE ).to(model_output.device ) UpperCamelCase__ : Tuple = (self._get_variance(SCREAMING_SNAKE_CASE , predicted_variance=SCREAMING_SNAKE_CASE ) ** 0.5) * noise UpperCamelCase__ : Dict = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return DDPMSchedulerOutput(prev_sample=SCREAMING_SNAKE_CASE , pred_original_sample=SCREAMING_SNAKE_CASE ) class __magic_name__ ( __lowerCAmelCase): def __init__( self : List[Any] , lowerCamelCase__ : UNetaDConditionModel , lowerCamelCase__ : Union[DDIMScheduler, DDPMScheduler] , lowerCamelCase__ : Optional[float] = 1.0 , ) -> Union[str, Any]: '''simple docstring''' super().__init__() UpperCamelCase__ : Optional[Any] = bit_scale UpperCamelCase__ : str = ( ddim_bit_scheduler_step if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else ddpm_bit_scheduler_step ) self.register_modules(unet=lowerCamelCase__ , scheduler=lowerCamelCase__ ) @torch.no_grad() def __call__( self : Tuple , lowerCamelCase__ : Optional[int] = 256 , lowerCamelCase__ : Optional[int] = 256 , lowerCamelCase__ : Optional[int] = 50 , lowerCamelCase__ : Optional[torch.Generator] = None , lowerCamelCase__ : Optional[int] = 1 , lowerCamelCase__ : Optional[str] = "pil" , lowerCamelCase__ : bool = True , **lowerCamelCase__ : Tuple , ) -> Union[Tuple, ImagePipelineOutput]: '''simple docstring''' UpperCamelCase__ : Tuple = torch.randn( (batch_size, self.unet.config.in_channels, height, width) , generator=lowerCamelCase__ , ) UpperCamelCase__ : Any = decimal_to_bits(lowerCamelCase__ ) * self.bit_scale UpperCamelCase__ : int = latents.to(self.device ) self.scheduler.set_timesteps(lowerCamelCase__ ) for t in self.progress_bar(self.scheduler.timesteps ): # predict the noise residual UpperCamelCase__ : List[Any] = self.unet(lowerCamelCase__ , lowerCamelCase__ ).sample # compute the previous noisy sample x_t -> x_t-1 UpperCamelCase__ : Dict = self.scheduler.step(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ).prev_sample UpperCamelCase__ : str = bits_to_decimal(lowerCamelCase__ ) if output_type == "pil": UpperCamelCase__ : List[str] = self.numpy_to_pil(lowerCamelCase__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCamelCase__ )
51
import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DonutImageProcessor class __magic_name__ ( unittest.TestCase): def __init__( self : Optional[Any] , lowerCamelCase__ : Any , lowerCamelCase__ : Tuple=7 , lowerCamelCase__ : List[Any]=3 , lowerCamelCase__ : Optional[int]=18 , lowerCamelCase__ : Any=30 , lowerCamelCase__ : int=400 , lowerCamelCase__ : List[str]=True , lowerCamelCase__ : Optional[Any]=None , lowerCamelCase__ : Union[str, Any]=True , lowerCamelCase__ : int=False , lowerCamelCase__ : Union[str, Any]=True , lowerCamelCase__ : int=True , lowerCamelCase__ : Dict=[0.5, 0.5, 0.5] , lowerCamelCase__ : str=[0.5, 0.5, 0.5] , ) -> Optional[Any]: '''simple docstring''' UpperCamelCase__ : Optional[Any] = parent UpperCamelCase__ : Dict = batch_size UpperCamelCase__ : List[Any] = num_channels UpperCamelCase__ : int = image_size UpperCamelCase__ : str = min_resolution UpperCamelCase__ : str = max_resolution UpperCamelCase__ : Tuple = do_resize UpperCamelCase__ : str = size if size is not None else {'''height''': 18, '''width''': 20} UpperCamelCase__ : Optional[Any] = do_thumbnail UpperCamelCase__ : int = do_align_axis UpperCamelCase__ : List[Any] = do_pad UpperCamelCase__ : List[Any] = do_normalize UpperCamelCase__ : Dict = image_mean UpperCamelCase__ : List[Any] = image_std def UpperCAmelCase__ ( self : List[Any] ) -> Any: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_thumbnail": self.do_thumbnail, "do_align_long_axis": self.do_align_axis, "do_pad": self.do_pad, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class __magic_name__ ( __lowerCAmelCase , unittest.TestCase): A: Tuple = DonutImageProcessor if is_vision_available() else None def UpperCAmelCase__ ( self : str ) -> int: '''simple docstring''' UpperCamelCase__ : int = DonutImageProcessingTester(self ) @property def UpperCAmelCase__ ( self : Tuple ) -> Optional[int]: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase__ ( self : Any ) -> Optional[Any]: '''simple docstring''' UpperCamelCase__ : Tuple = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase__ , '''do_resize''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''size''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''do_thumbnail''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''do_align_long_axis''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''do_pad''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''do_normalize''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''image_mean''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''image_std''' ) ) def UpperCAmelCase__ ( self : Optional[Any] ) -> Tuple: '''simple docstring''' UpperCamelCase__ : Union[str, Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''height''': 18, '''width''': 20} ) UpperCamelCase__ : Any = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {'''height''': 42, '''width''': 42} ) # Previous config had dimensions in (width, height) order UpperCamelCase__ : Dict = self.image_processing_class.from_dict(self.image_processor_dict , size=(42, 84) ) self.assertEqual(image_processor.size , {'''height''': 84, '''width''': 42} ) def UpperCAmelCase__ ( self : Any ) -> str: '''simple docstring''' pass @is_flaky() def UpperCAmelCase__ ( self : Optional[int] ) -> Any: '''simple docstring''' UpperCamelCase__ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCamelCase__ : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ , Image.Image ) # Test not batched input UpperCamelCase__ : Dict = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched UpperCamelCase__ : List[str] = image_processing(lowerCamelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) @is_flaky() def UpperCAmelCase__ ( self : Optional[int] ) -> List[str]: '''simple docstring''' UpperCamelCase__ : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCamelCase__ : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase__ , numpify=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ , np.ndarray ) # Test not batched input UpperCamelCase__ : List[str] = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched UpperCamelCase__ : List[Any] = image_processing(lowerCamelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) @is_flaky() def UpperCAmelCase__ ( self : str ) -> Tuple: '''simple docstring''' UpperCamelCase__ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCamelCase__ : List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase__ , torchify=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ , torch.Tensor ) # Test not batched input UpperCamelCase__ : str = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched UpperCamelCase__ : List[str] = image_processing(lowerCamelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , )
51
1
"""simple docstring""" from math import pi def UpperCAmelCase__ (snake_case__ : int , snake_case__ : int ): """simple docstring""" return 2 * pi * radius * (angle / 3_60) if __name__ == "__main__": print(arc_length(90, 10))
64
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : Union[str, Any] ): """simple docstring""" stooge(snake_case__ , 0 , len(snake_case__ ) - 1 ) return arr def UpperCAmelCase__ (snake_case__ : List[Any] , snake_case__ : Any , snake_case__ : int ): """simple docstring""" if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: _snake_case , _snake_case : Tuple = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: _snake_case : Dict = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(snake_case__ , snake_case__ , (h - t) ) # Recursively sort last 2/3 elements stooge(snake_case__ , i + t , (snake_case__) ) # Recursively sort first 2/3 elements stooge(snake_case__ , snake_case__ , (h - t) ) if __name__ == "__main__": A_ = input('''Enter numbers separated by a comma:\n''').strip() A_ = [int(item) for item in user_input.split(''',''')] print(stooge_sort(unsorted))
64
1
from argparse import ArgumentParser from ..pipelines import Pipeline, PipelineDataFormat, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand lowerCAmelCase__ = logging.get_logger(__name__) # pylint: disable=invalid-name def a__ ( SCREAMING_SNAKE_CASE : str ): '''simple docstring''' if not path: return "pipe" for ext in PipelineDataFormat.SUPPORTED_FORMATS: if path.endswith(SCREAMING_SNAKE_CASE ): return ext raise Exception( f"""Unable to determine file format from file extension {path}. """ f"""Please provide the format through --format {PipelineDataFormat.SUPPORTED_FORMATS}""" ) def a__ ( SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' lowerCAmelCase : Any = pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) lowerCAmelCase : List[str] = try_infer_format_from_ext(args.input ) if args.format == "infer" else args.format lowerCAmelCase : Union[str, Any] = PipelineDataFormat.from_str( format=SCREAMING_SNAKE_CASE , output_path=args.output , input_path=args.input , column=args.column if args.column else nlp.default_input_names , overwrite=args.overwrite , ) return RunCommand(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) class SCREAMING_SNAKE_CASE__ ( lowercase ): """simple docstring""" def __init__( self , snake_case__ , snake_case__ ): """simple docstring""" lowerCAmelCase : Optional[Any] = nlp lowerCAmelCase : List[Any] = reader @staticmethod def lowercase__ ( snake_case__ ): """simple docstring""" lowerCAmelCase : str = parser.add_parser("run" , help="Run a pipeline through the CLI" ) run_parser.add_argument("--task" , choices=get_supported_tasks() , help="Task to run" ) run_parser.add_argument("--input" , type=snake_case__ , help="Path to the file to use for inference" ) run_parser.add_argument("--output" , type=snake_case__ , help="Path to the file that will be used post to write results." ) run_parser.add_argument("--model" , type=snake_case__ , help="Name or path to the model to instantiate." ) run_parser.add_argument("--config" , type=snake_case__ , help="Name or path to the model's config to instantiate." ) run_parser.add_argument( "--tokenizer" , type=snake_case__ , help="Name of the tokenizer to use. (default: same as the model name)" ) run_parser.add_argument( "--column" , type=snake_case__ , help="Name of the column to use as input. (For multi columns input as QA use column1,columns2)" , ) run_parser.add_argument( "--format" , type=snake_case__ , default="infer" , choices=PipelineDataFormat.SUPPORTED_FORMATS , help="Input format to read from" , ) run_parser.add_argument( "--device" , type=snake_case__ , default=-1 , help="Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)" , ) run_parser.add_argument("--overwrite" , action="store_true" , help="Allow overwriting the output file." ) run_parser.set_defaults(func=snake_case__ ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : str = self._nlp, [] for entry in self._reader: lowerCAmelCase : Any = nlp(**snake_case__ ) if self._reader.is_multi_columns else nlp(snake_case__ ) if isinstance(snake_case__ , snake_case__ ): outputs.append(snake_case__ ) else: outputs += output # Saving data if self._nlp.binary_output: lowerCAmelCase : Any = self._reader.save_binary(snake_case__ ) logger.warning(f"""Current pipeline requires output to be in binary format, saving at {binary_path}""" ) else: self._reader.save(snake_case__ )
361
"""simple docstring""" from typing import List import jiwer import jiwer.transforms as tr from packaging import version import datasets from datasets.config import PY_VERSION if PY_VERSION < version.parse('''3.8'''): import importlib_metadata else: import importlib.metadata as importlib_metadata lowerCAmelCase__ = '''''' if version.parse(importlib_metadata.version('''jiwer''')) < version.parse('''2.3.0'''): class SCREAMING_SNAKE_CASE__ ( tr.AbstractTransform ): """simple docstring""" def __init__( self , snake_case__ = " " ): """simple docstring""" lowerCAmelCase : List[Any] = sentence_delimiter def lowercase__ ( self , snake_case__ ): """simple docstring""" return list(snake_case__ ) def lowercase__ ( self , snake_case__ ): """simple docstring""" lowerCAmelCase : Any = [] for sent_idx, sentence in enumerate(snake_case__ ): chars.extend(self.process_string(snake_case__ ) ) if self.sentence_delimiter is not None and self.sentence_delimiter != "" and sent_idx < len(snake_case__ ) - 1: chars.append(self.sentence_delimiter ) return chars lowerCAmelCase__ = tr.Compose( [tr.RemoveMultipleSpaces(), tr.Strip(), SentencesToListOfCharacters(SENTENCE_DELIMITER)] ) else: lowerCAmelCase__ = tr.Compose( [ tr.RemoveMultipleSpaces(), tr.Strip(), tr.ReduceToSingleSentence(SENTENCE_DELIMITER), tr.ReduceToListOfListOfChars(), ] ) lowerCAmelCase__ = '''\ @inproceedings{inproceedings, author = {Morris, Andrew and Maier, Viktoria and Green, Phil}, year = {2004}, month = {01}, pages = {}, title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.} } ''' lowerCAmelCase__ = '''\ Character error rate (CER) is a common metric of the performance of an automatic speech recognition system. CER is similar to Word Error Rate (WER), but operates on character instead of word. Please refer to docs of WER for further information. Character error rate can be computed as: CER = (S + D + I) / N = (S + D + I) / (S + D + C) where S is the number of substitutions, D is the number of deletions, I is the number of insertions, C is the number of correct characters, N is the number of characters in the reference (N=S+D+C). CER\'s output is not always a number between 0 and 1, in particular when there is a high number of insertions. This value is often associated to the percentage of characters that were incorrectly predicted. The lower the value, the better the performance of the ASR system with a CER of 0 being a perfect score. ''' lowerCAmelCase__ = ''' Computes CER score of transcribed segments against references. Args: references: list of references for each speech input. predictions: list of transcribtions to score. concatenate_texts: Whether or not to concatenate sentences before evaluation, set to True for more accurate result. Returns: (float): the character error rate Examples: >>> predictions = ["this is the prediction", "there is an other sample"] >>> references = ["this is the reference", "there is another one"] >>> cer = datasets.load_metric("cer") >>> cer_score = cer.compute(predictions=predictions, references=references) >>> print(cer_score) 0.34146341463414637 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class SCREAMING_SNAKE_CASE__ ( datasets.Metric ): """simple docstring""" def lowercase__ ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/jitsi/jiwer/"] , reference_urls=[ "https://en.wikipedia.org/wiki/Word_error_rate", "https://sites.google.com/site/textdigitisation/qualitymeasures/computingerrorrates", ] , ) def lowercase__ ( self , snake_case__ , snake_case__ , snake_case__=False ): """simple docstring""" if concatenate_texts: return jiwer.compute_measures( snake_case__ , snake_case__ , truth_transform=snake_case__ , hypothesis_transform=snake_case__ , )["wer"] lowerCAmelCase : List[Any] = 0 lowerCAmelCase : Optional[Any] = 0 for prediction, reference in zip(snake_case__ , snake_case__ ): lowerCAmelCase : Optional[int] = jiwer.compute_measures( snake_case__ , snake_case__ , truth_transform=snake_case__ , hypothesis_transform=snake_case__ , ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
133
0