code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
import argparse import json import torch from diffusers import DDPMScheduler, LDMPipeline, UNetaDModel, VQModel def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=1 ): '''simple docstring''' if n_shave_prefix_segments >= 0: return ".".join(path.split('''.''' )[n_shave_prefix_segments:] ) else: return ".".join(path.split('''.''' )[:n_shave_prefix_segments] ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=0 ): '''simple docstring''' lowercase = [] for old_item in old_list: lowercase = old_item.replace('''in_layers.0''' , '''norm1''' ) lowercase = new_item.replace('''in_layers.2''' , '''conv1''' ) lowercase = new_item.replace('''out_layers.0''' , '''norm2''' ) lowercase = new_item.replace('''out_layers.3''' , '''conv2''' ) lowercase = new_item.replace('''emb_layers.1''' , '''time_emb_proj''' ) lowercase = new_item.replace('''skip_connection''' , '''conv_shortcut''' ) lowercase = shave_segments(lowerCAmelCase__ , n_shave_prefix_segments=lowerCAmelCase__ ) mapping.append({'''old''': old_item, '''new''': new_item} ) return mapping def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=0 ): '''simple docstring''' lowercase = [] for old_item in old_list: lowercase = old_item lowercase = new_item.replace('''norm.weight''' , '''group_norm.weight''' ) lowercase = new_item.replace('''norm.bias''' , '''group_norm.bias''' ) lowercase = new_item.replace('''proj_out.weight''' , '''proj_attn.weight''' ) lowercase = new_item.replace('''proj_out.bias''' , '''proj_attn.bias''' ) lowercase = shave_segments(lowerCAmelCase__ , n_shave_prefix_segments=lowerCAmelCase__ ) mapping.append({'''old''': old_item, '''new''': new_item} ) return mapping def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None ): '''simple docstring''' assert isinstance(lowerCAmelCase__ , lowerCAmelCase__ ), "Paths should be a list of dicts containing 'old' and 'new' keys." # Splits the attention layers into three variables. if attention_paths_to_split is not None: for path, path_map in attention_paths_to_split.items(): lowercase = old_checkpoint[path] lowercase = old_tensor.shape[0] // 3 lowercase = (-1, channels) if len(old_tensor.shape ) == 3 else (-1) lowercase = old_tensor.shape[0] // config['''num_head_channels'''] // 3 lowercase = old_tensor.reshape((num_heads, 3 * channels // num_heads) + old_tensor.shape[1:] ) lowercase , lowercase , lowercase = old_tensor.split(channels // num_heads , dim=1 ) lowercase = query.reshape(lowerCAmelCase__ ) lowercase = key.reshape(lowerCAmelCase__ ) lowercase = value.reshape(lowerCAmelCase__ ) for path in paths: lowercase = path['''new'''] # These have already been assigned if attention_paths_to_split is not None and new_path in attention_paths_to_split: continue # Global renaming happens here lowercase = new_path.replace('''middle_block.0''' , '''mid_block.resnets.0''' ) lowercase = new_path.replace('''middle_block.1''' , '''mid_block.attentions.0''' ) lowercase = new_path.replace('''middle_block.2''' , '''mid_block.resnets.1''' ) if additional_replacements is not None: for replacement in additional_replacements: lowercase = new_path.replace(replacement['''old'''] , replacement['''new'''] ) # proj_attn.weight has to be converted from conv 1D to linear if "proj_attn.weight" in new_path: lowercase = old_checkpoint[path['''old''']][:, :, 0] else: lowercase = old_checkpoint[path['''old''']] def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = {} lowercase = checkpoint['''time_embed.0.weight'''] lowercase = checkpoint['''time_embed.0.bias'''] lowercase = checkpoint['''time_embed.2.weight'''] lowercase = checkpoint['''time_embed.2.bias'''] lowercase = checkpoint['''input_blocks.0.0.weight'''] lowercase = checkpoint['''input_blocks.0.0.bias'''] lowercase = checkpoint['''out.0.weight'''] lowercase = checkpoint['''out.0.bias'''] lowercase = checkpoint['''out.2.weight'''] lowercase = checkpoint['''out.2.bias'''] # Retrieves the keys for the input blocks only lowercase = len({'''.'''.join(layer.split('''.''' )[:2] ) for layer in checkpoint if '''input_blocks''' in layer} ) lowercase = { layer_id: [key for key in checkpoint if f'input_blocks.{layer_id}' in key] for layer_id in range(lowerCAmelCase__ ) } # Retrieves the keys for the middle blocks only lowercase = len({'''.'''.join(layer.split('''.''' )[:2] ) for layer in checkpoint if '''middle_block''' in layer} ) lowercase = { layer_id: [key for key in checkpoint if f'middle_block.{layer_id}' in key] for layer_id in range(lowerCAmelCase__ ) } # Retrieves the keys for the output blocks only lowercase = len({'''.'''.join(layer.split('''.''' )[:2] ) for layer in checkpoint if '''output_blocks''' in layer} ) lowercase = { layer_id: [key for key in checkpoint if f'output_blocks.{layer_id}' in key] for layer_id in range(lowerCAmelCase__ ) } for i in range(1 , lowerCAmelCase__ ): lowercase = (i - 1) // (config['''num_res_blocks'''] + 1) lowercase = (i - 1) % (config['''num_res_blocks'''] + 1) lowercase = [key for key in input_blocks[i] if f'input_blocks.{i}.0' in key] lowercase = [key for key in input_blocks[i] if f'input_blocks.{i}.1' in key] if f'input_blocks.{i}.0.op.weight' in checkpoint: lowercase = checkpoint[ f'input_blocks.{i}.0.op.weight' ] lowercase = checkpoint[ f'input_blocks.{i}.0.op.bias' ] continue lowercase = renew_resnet_paths(lowerCAmelCase__ ) lowercase = {'''old''': f'input_blocks.{i}.0', '''new''': f'down_blocks.{block_id}.resnets.{layer_in_block_id}'} lowercase = {'''old''': '''resnets.2.op''', '''new''': '''downsamplers.0.op'''} assign_to_checkpoint( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , additional_replacements=[meta_path, resnet_op] , config=lowerCAmelCase__ ) if len(lowerCAmelCase__ ): lowercase = renew_attention_paths(lowerCAmelCase__ ) lowercase = { '''old''': f'input_blocks.{i}.1', '''new''': f'down_blocks.{block_id}.attentions.{layer_in_block_id}', } lowercase = { f'input_blocks.{i}.1.qkv.bias': { '''key''': f'down_blocks.{block_id}.attentions.{layer_in_block_id}.key.bias', '''query''': f'down_blocks.{block_id}.attentions.{layer_in_block_id}.query.bias', '''value''': f'down_blocks.{block_id}.attentions.{layer_in_block_id}.value.bias', }, f'input_blocks.{i}.1.qkv.weight': { '''key''': f'down_blocks.{block_id}.attentions.{layer_in_block_id}.key.weight', '''query''': f'down_blocks.{block_id}.attentions.{layer_in_block_id}.query.weight', '''value''': f'down_blocks.{block_id}.attentions.{layer_in_block_id}.value.weight', }, } assign_to_checkpoint( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , additional_replacements=[meta_path] , attention_paths_to_split=lowerCAmelCase__ , config=lowerCAmelCase__ , ) lowercase = middle_blocks[0] lowercase = middle_blocks[1] lowercase = middle_blocks[2] lowercase = renew_resnet_paths(lowerCAmelCase__ ) assign_to_checkpoint(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , config=lowerCAmelCase__ ) lowercase = renew_resnet_paths(lowerCAmelCase__ ) assign_to_checkpoint(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , config=lowerCAmelCase__ ) lowercase = renew_attention_paths(lowerCAmelCase__ ) lowercase = { '''middle_block.1.qkv.bias''': { '''key''': '''mid_block.attentions.0.key.bias''', '''query''': '''mid_block.attentions.0.query.bias''', '''value''': '''mid_block.attentions.0.value.bias''', }, '''middle_block.1.qkv.weight''': { '''key''': '''mid_block.attentions.0.key.weight''', '''query''': '''mid_block.attentions.0.query.weight''', '''value''': '''mid_block.attentions.0.value.weight''', }, } assign_to_checkpoint( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , attention_paths_to_split=lowerCAmelCase__ , config=lowerCAmelCase__ ) for i in range(lowerCAmelCase__ ): lowercase = i // (config['''num_res_blocks'''] + 1) lowercase = i % (config['''num_res_blocks'''] + 1) lowercase = [shave_segments(lowerCAmelCase__ , 2 ) for name in output_blocks[i]] lowercase = {} for layer in output_block_layers: lowercase , lowercase = layer.split('''.''' )[0], shave_segments(lowerCAmelCase__ , 1 ) if layer_id in output_block_list: output_block_list[layer_id].append(lowerCAmelCase__ ) else: lowercase = [layer_name] if len(lowerCAmelCase__ ) > 1: lowercase = [key for key in output_blocks[i] if f'output_blocks.{i}.0' in key] lowercase = [key for key in output_blocks[i] if f'output_blocks.{i}.1' in key] lowercase = renew_resnet_paths(lowerCAmelCase__ ) lowercase = renew_resnet_paths(lowerCAmelCase__ ) lowercase = {'''old''': f'output_blocks.{i}.0', '''new''': f'up_blocks.{block_id}.resnets.{layer_in_block_id}'} assign_to_checkpoint(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , additional_replacements=[meta_path] , config=lowerCAmelCase__ ) if ["conv.weight", "conv.bias"] in output_block_list.values(): lowercase = list(output_block_list.values() ).index(['''conv.weight''', '''conv.bias'''] ) lowercase = checkpoint[ f'output_blocks.{i}.{index}.conv.weight' ] lowercase = checkpoint[ f'output_blocks.{i}.{index}.conv.bias' ] # Clear attentions as they have been attributed above. if len(lowerCAmelCase__ ) == 2: lowercase = [] if len(lowerCAmelCase__ ): lowercase = renew_attention_paths(lowerCAmelCase__ ) lowercase = { '''old''': f'output_blocks.{i}.1', '''new''': f'up_blocks.{block_id}.attentions.{layer_in_block_id}', } lowercase = { f'output_blocks.{i}.1.qkv.bias': { '''key''': f'up_blocks.{block_id}.attentions.{layer_in_block_id}.key.bias', '''query''': f'up_blocks.{block_id}.attentions.{layer_in_block_id}.query.bias', '''value''': f'up_blocks.{block_id}.attentions.{layer_in_block_id}.value.bias', }, f'output_blocks.{i}.1.qkv.weight': { '''key''': f'up_blocks.{block_id}.attentions.{layer_in_block_id}.key.weight', '''query''': f'up_blocks.{block_id}.attentions.{layer_in_block_id}.query.weight', '''value''': f'up_blocks.{block_id}.attentions.{layer_in_block_id}.value.weight', }, } assign_to_checkpoint( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , additional_replacements=[meta_path] , attention_paths_to_split=to_split if any('''qkv''' in key for key in attentions ) else None , config=lowerCAmelCase__ , ) else: lowercase = renew_resnet_paths(lowerCAmelCase__ , n_shave_prefix_segments=1 ) for path in resnet_0_paths: lowercase = '''.'''.join(['''output_blocks''', str(lowerCAmelCase__ ), path['''old''']] ) lowercase = '''.'''.join(['''up_blocks''', str(lowerCAmelCase__ ), '''resnets''', str(lowerCAmelCase__ ), path['''new''']] ) lowercase = checkpoint[old_path] return new_checkpoint if __name__ == "__main__": lowercase__ :int = argparse.ArgumentParser() parser.add_argument( "--checkpoint_path", default=None, type=str, required=True, help="Path to the checkpoint to convert." ) parser.add_argument( "--config_file", default=None, type=str, required=True, help="The config json file corresponding to the architecture.", ) parser.add_argument("--dump_path", default=None, type=str, required=True, help="Path to the output model.") lowercase__ :Optional[int] = parser.parse_args() lowercase__ :List[Any] = torch.load(args.checkpoint_path) with open(args.config_file) as f: lowercase__ :Union[str, Any] = json.loads(f.read()) lowercase__ :List[Any] = convert_ldm_checkpoint(checkpoint, config) if "ldm" in config: del config["ldm"] lowercase__ :List[Any] = UNetaDModel(**config) model.load_state_dict(converted_checkpoint) try: lowercase__ :int = DDPMScheduler.from_config("/".join(args.checkpoint_path.split("/")[:-1])) lowercase__ :List[Any] = VQModel.from_pretrained("/".join(args.checkpoint_path.split("/")[:-1])) lowercase__ :Optional[int] = LDMPipeline(unet=model, scheduler=scheduler, vae=vqvae) pipe.save_pretrained(args.dump_path) except: # noqa: E722 model.save_pretrained(args.dump_path)
633
import random import unittest import numpy as np import transformers from transformers import is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax if is_flax_available(): import os import jax.numpy as jnp from jax import jit from transformers import AutoTokenizer, FlaxAutoModelForCausalLM from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model lowercase__ :Optional[Any] = "0.12" # assumed parallelism: 8 if is_torch_available(): import torch def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None ): '''simple docstring''' if rng is None: lowercase = random.Random() lowercase = 1 for dim in shape: total_dims *= dim lowercase = [] for _ in range(lowerCAmelCase__ ): values.append(rng.randint(0 , vocab_size - 1 ) ) lowercase = np.array(lowerCAmelCase__ , dtype=jnp.intaa ).reshape(lowerCAmelCase__ ) return output def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=None ): '''simple docstring''' lowercase = ids_tensor(lowerCAmelCase__ , vocab_size=2 , rng=lowerCAmelCase__ ) # make sure that at least one token is attended to for each batch lowercase = 1 return attn_mask @require_flax class lowercase : lowercase_ : Any =None lowercase_ : List[str] =() def A__ ( self): lowercase , lowercase = self.model_tester.prepare_config_and_inputs_for_common() # cut to half length & take max batch_size 3 lowercase = 2 lowercase = inputs['''input_ids'''].shape[-1] // 2 lowercase = inputs['''input_ids'''][:max_batch_size, :sequence_length] lowercase = jnp.ones_like(A__) lowercase = attention_mask[:max_batch_size, :sequence_length] # generate max 5 tokens lowercase = input_ids.shape[-1] + 5 if config.eos_token_id is not None and config.pad_token_id is None: # hack to allow generate for models such as GPT2 as is done in `generate()` lowercase = config.eos_token_id return config, input_ids, attention_mask, max_length @is_pt_flax_cross_test def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = False lowercase = max_length lowercase = 0 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model_class.__name__[4:] # Skip the "Flax" at the beginning lowercase = getattr(A__ ,A__) lowercase = pt_model_class(A__).eval() lowercase = load_flax_weights_in_pytorch_model(A__ ,flax_model.params) lowercase = flax_model.generate(A__).sequences lowercase = pt_model.generate(torch.tensor(A__ ,dtype=torch.long)) if flax_generation_outputs.shape[-1] > pt_generation_outputs.shape[-1]: lowercase = flax_generation_outputs[:, : pt_generation_outputs.shape[-1]] self.assertListEqual(pt_generation_outputs.numpy().tolist() ,flax_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = False lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = True lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = False lowercase = max_length lowercase = 2 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = False lowercase = max_length lowercase = 2 lowercase = 2 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[0] ,input_ids.shape[0] * config.num_return_sequences) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = True lowercase = max_length lowercase = 0.8 lowercase = 1_0 lowercase = 0.3 lowercase = 1 lowercase = 8 lowercase = 9 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = max_length lowercase = 1 lowercase = 8 lowercase = 9 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = max_length lowercase = 2 lowercase = 1 lowercase = 8 lowercase = 9 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() # pad attention mask on the left lowercase = attention_mask.at[(0, 0)].set(0) lowercase = False lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__ ,attention_mask=A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__ ,attention_mask=A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() # pad attention mask on the left lowercase = attention_mask.at[(0, 0)].set(0) lowercase = True lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__ ,attention_mask=A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__ ,attention_mask=A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() # pad attention mask on the left lowercase = attention_mask.at[(0, 0)].set(0) lowercase = 2 lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__ ,attention_mask=A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__ ,attention_mask=A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) @require_flax class lowercase ( unittest.TestCase ): def A__ ( self): lowercase = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-bert''') lowercase = FlaxAutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-bert-flax-only''') lowercase = '''Hello world''' lowercase = tokenizer(A__ ,return_tensors='''np''').input_ids # typos are quickly detected (the correct argument is `do_sample`) with self.assertRaisesRegex(A__ ,'''do_samples'''): model.generate(A__ ,do_samples=A__) # arbitrary arguments that will not be used anywhere are also not accepted with self.assertRaisesRegex(A__ ,'''foo'''): lowercase = {'''foo''': '''bar'''} model.generate(A__ ,**A__)
633
1
import sys import turtle def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' return (pa[0] + pa[0]) / 2, (pa[1] + pa[1]) / 2 def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ): '''simple docstring''' my_pen.up() my_pen.goto(vertexa[0] , vertexa[1] ) my_pen.down() my_pen.goto(vertexa[0] , vertexa[1] ) my_pen.goto(vertexa[0] , vertexa[1] ) my_pen.goto(vertexa[0] , vertexa[1] ) if depth == 0: return triangle(lowerCAmelCase__ , get_mid(lowerCAmelCase__ , lowerCAmelCase__ ) , get_mid(lowerCAmelCase__ , lowerCAmelCase__ ) , depth - 1 ) triangle(lowerCAmelCase__ , get_mid(lowerCAmelCase__ , lowerCAmelCase__ ) , get_mid(lowerCAmelCase__ , lowerCAmelCase__ ) , depth - 1 ) triangle(lowerCAmelCase__ , get_mid(lowerCAmelCase__ , lowerCAmelCase__ ) , get_mid(lowerCAmelCase__ , lowerCAmelCase__ ) , depth - 1 ) if __name__ == "__main__": if len(sys.argv) != 2: raise ValueError( "Correct format for using this script: " "python fractals.py <int:depth_for_fractal>" ) lowercase__ :Dict = turtle.Turtle() my_pen.ht() my_pen.speed(5) my_pen.pencolor("red") lowercase__ :Optional[int] = [(-175, -125), (0, 175), (175, -125)] # vertices of triangle triangle(vertices[0], vertices[1], vertices[2], int(sys.argv[1]))
633
import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : List[str] =(UniPCMultistepScheduler,) lowercase_ : Tuple =(('''num_inference_steps''', 25),) def A__ ( self ,**A__): lowercase = { '''num_train_timesteps''': 1_0_0_0, '''beta_start''': 0.0001, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', '''solver_order''': 2, '''solver_type''': '''bh2''', } config.update(**A__) return config def A__ ( self ,A__=0 ,**A__): lowercase = dict(self.forward_default_kwargs) lowercase = kwargs.pop('''num_inference_steps''' ,A__) lowercase = self.dummy_sample lowercase = 0.1 * sample lowercase = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) scheduler.set_timesteps(A__) # copy over dummy past residuals lowercase = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A__) lowercase = scheduler_class.from_pretrained(A__) new_scheduler.set_timesteps(A__) # copy over dummy past residuals lowercase = dummy_past_residuals[: new_scheduler.config.solver_order] lowercase , lowercase = sample, sample for t in range(A__ ,time_step + scheduler.config.solver_order + 1): lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = new_scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1E-5, "Scheduler outputs are not identical" def A__ ( self ,A__=0 ,**A__): lowercase = dict(self.forward_default_kwargs) lowercase = kwargs.pop('''num_inference_steps''' ,A__) lowercase = self.dummy_sample lowercase = 0.1 * sample lowercase = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config() lowercase = scheduler_class(**A__) scheduler.set_timesteps(A__) # copy over dummy past residuals (must be after setting timesteps) lowercase = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A__) lowercase = scheduler_class.from_pretrained(A__) # copy over dummy past residuals new_scheduler.set_timesteps(A__) # copy over dummy past residual (must be after setting timesteps) lowercase = dummy_past_residuals[: new_scheduler.config.solver_order] lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = new_scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1E-5, "Scheduler outputs are not identical" def A__ ( self ,A__=None ,**A__): if scheduler is None: lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) lowercase = 1_0 lowercase = self.dummy_model() lowercase = self.dummy_sample_deter scheduler.set_timesteps(A__) for i, t in enumerate(scheduler.timesteps): lowercase = model(A__ ,A__) lowercase = scheduler.step(A__ ,A__ ,A__).prev_sample return sample def A__ ( self): lowercase = dict(self.forward_default_kwargs) lowercase = kwargs.pop('''num_inference_steps''' ,A__) for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config() lowercase = scheduler_class(**A__) lowercase = self.dummy_sample lowercase = 0.1 * sample if num_inference_steps is not None and hasattr(A__ ,'''set_timesteps'''): scheduler.set_timesteps(A__) elif num_inference_steps is not None and not hasattr(A__ ,'''set_timesteps'''): lowercase = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) lowercase = [residual + 0.2, residual + 0.15, residual + 0.10] lowercase = dummy_past_residuals[: scheduler.config.solver_order] lowercase = scheduler.timesteps[5] lowercase = scheduler.timesteps[6] lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample self.assertEqual(output_a.shape ,sample.shape) self.assertEqual(output_a.shape ,output_a.shape) def A__ ( self): # make sure that iterating over schedulers with same config names gives same results # for defaults lowercase = UniPCMultistepScheduler(**self.get_scheduler_config()) lowercase = self.full_loop(scheduler=A__) lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.2464) < 1E-3 lowercase = DPMSolverSinglestepScheduler.from_config(scheduler.config) lowercase = DEISMultistepScheduler.from_config(scheduler.config) lowercase = DPMSolverMultistepScheduler.from_config(scheduler.config) lowercase = UniPCMultistepScheduler.from_config(scheduler.config) lowercase = self.full_loop(scheduler=A__) lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.2464) < 1E-3 def A__ ( self): for timesteps in [2_5, 5_0, 1_0_0, 9_9_9, 1_0_0_0]: self.check_over_configs(num_train_timesteps=A__) def A__ ( self): self.check_over_configs(thresholding=A__) for order in [1, 2, 3]: for solver_type in ["bh1", "bh2"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=A__ ,prediction_type=A__ ,sample_max_value=A__ ,solver_order=A__ ,solver_type=A__ ,) def A__ ( self): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=A__) def A__ ( self): for solver_type in ["bh1", "bh2"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=A__ ,solver_type=A__ ,prediction_type=A__ ,) lowercase = self.full_loop( solver_order=A__ ,solver_type=A__ ,prediction_type=A__ ,) assert not torch.isnan(A__).any(), "Samples have nan numbers" def A__ ( self): self.check_over_configs(lower_order_final=A__) self.check_over_configs(lower_order_final=A__) def A__ ( self): for num_inference_steps in [1, 2, 3, 5, 1_0, 5_0, 1_0_0, 9_9_9, 1_0_0_0]: self.check_over_forward(num_inference_steps=A__ ,time_step=0) def A__ ( self): lowercase = self.full_loop() lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.2464) < 1E-3 def A__ ( self): lowercase = self.full_loop(prediction_type='''v_prediction''') lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.1014) < 1E-3 def A__ ( self): lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config(thresholding=A__ ,dynamic_thresholding_ratio=0) lowercase = scheduler_class(**A__) lowercase = 1_0 lowercase = self.dummy_model() lowercase = self.dummy_sample_deter.half() scheduler.set_timesteps(A__) for i, t in enumerate(scheduler.timesteps): lowercase = model(A__ ,A__) lowercase = scheduler.step(A__ ,A__ ,A__).prev_sample assert sample.dtype == torch.floataa def A__ ( self ,**A__): for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) scheduler.set_timesteps(scheduler.config.num_train_timesteps) assert len(scheduler.timesteps.unique()) == scheduler.num_inference_steps
633
1
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging lowercase__ :Tuple = logging.get_logger(__name__) lowercase__ :str = {"vocab_file": "spiece.model"} lowercase__ :List[str] = { "vocab_file": { "albert-base-v1": "https://huggingface.co/albert-base-v1/resolve/main/spiece.model", "albert-large-v1": "https://huggingface.co/albert-large-v1/resolve/main/spiece.model", "albert-xlarge-v1": "https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model", "albert-xxlarge-v1": "https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model", "albert-base-v2": "https://huggingface.co/albert-base-v2/resolve/main/spiece.model", "albert-large-v2": "https://huggingface.co/albert-large-v2/resolve/main/spiece.model", "albert-xlarge-v2": "https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model", "albert-xxlarge-v2": "https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model", } } lowercase__ :int = { "albert-base-v1": 512, "albert-large-v1": 512, "albert-xlarge-v1": 512, "albert-xxlarge-v1": 512, "albert-base-v2": 512, "albert-large-v2": 512, "albert-xlarge-v2": 512, "albert-xxlarge-v2": 512, } lowercase__ :Optional[Any] = "▁" class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Optional[int] =VOCAB_FILES_NAMES lowercase_ : List[str] =PRETRAINED_VOCAB_FILES_MAP lowercase_ : Optional[Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self ,A__ ,A__=True ,A__=True ,A__=False ,A__="[CLS]" ,A__="[SEP]" ,A__="<unk>" ,A__="[SEP]" ,A__="<pad>" ,A__="[CLS]" ,A__="[MASK]" ,A__ = None ,**A__ ,): # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. lowercase = ( AddedToken(A__ ,lstrip=A__ ,rstrip=A__ ,normalized=A__) if isinstance(A__ ,A__) else mask_token ) lowercase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=A__ ,remove_space=A__ ,keep_accents=A__ ,bos_token=A__ ,eos_token=A__ ,unk_token=A__ ,sep_token=A__ ,pad_token=A__ ,cls_token=A__ ,mask_token=A__ ,sp_model_kwargs=self.sp_model_kwargs ,**A__ ,) lowercase = do_lower_case lowercase = remove_space lowercase = keep_accents lowercase = vocab_file lowercase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(A__) @property def A__ ( self): return len(self.sp_model) def A__ ( self): lowercase = {self.convert_ids_to_tokens(A__): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __getstate__( self): lowercase = self.__dict__.copy() lowercase = None return state def __setstate__( self ,A__): lowercase = d # for backward compatibility if not hasattr(self ,'''sp_model_kwargs'''): lowercase = {} lowercase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def A__ ( self ,A__): if self.remove_space: lowercase = ''' '''.join(inputs.strip().split()) else: lowercase = inputs lowercase = outputs.replace('''``''' ,'''"''').replace('''\'\'''' ,'''"''') if not self.keep_accents: lowercase = unicodedata.normalize('''NFKD''' ,A__) lowercase = ''''''.join([c for c in outputs if not unicodedata.combining(A__)]) if self.do_lower_case: lowercase = outputs.lower() return outputs def A__ ( self ,A__): lowercase = self.preprocess_text(A__) lowercase = self.sp_model.encode(A__ ,out_type=A__) lowercase = [] for piece in pieces: if len(A__) > 1 and piece[-1] == str(''',''') and piece[-2].isdigit(): lowercase = self.sp_model.EncodeAsPieces(piece[:-1].replace(A__ ,'''''')) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0]) == 1: lowercase = cur_pieces[1:] else: lowercase = cur_pieces[0][1:] cur_pieces.append(piece[-1]) new_pieces.extend(A__) else: new_pieces.append(A__) return new_pieces def A__ ( self ,A__): return self.sp_model.PieceToId(A__) def A__ ( self ,A__): return self.sp_model.IdToPiece(A__) def A__ ( self ,A__): lowercase = [] lowercase = '''''' lowercase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(A__) + token lowercase = True lowercase = [] else: current_sub_tokens.append(A__) lowercase = False out_string += self.sp_model.decode(A__) return out_string.strip() def A__ ( self ,A__ ,A__ = None): lowercase = [self.sep_token_id] lowercase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def A__ ( self ,A__ ,A__ = None ,A__ = False): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=A__ ,token_ids_a=A__ ,already_has_special_tokens=A__) if token_ids_a is not None: return [1] + ([0] * len(A__)) + [1] + ([0] * len(A__)) + [1] return [1] + ([0] * len(A__)) + [1] def A__ ( self ,A__ ,A__ = None): lowercase = [self.sep_token_id] lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def A__ ( self ,A__ ,A__ = None): if not os.path.isdir(A__): logger.error(f'Vocabulary path ({save_directory}) should be a directory') return lowercase = os.path.join( A__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file''']) if os.path.abspath(self.vocab_file) != os.path.abspath(A__) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file ,A__) elif not os.path.isfile(self.vocab_file): with open(A__ ,'''wb''') as fi: lowercase = self.sp_model.serialized_model_proto() fi.write(A__) return (out_vocab_file,)
633
import argparse from collections import defaultdict import yaml lowercase__ :Optional[int] = "docs/source/en/_toctree.yml" def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = defaultdict(lowerCAmelCase__ ) for doc in model_doc: counts[doc["local"]] += 1 lowercase = [key for key, value in counts.items() if value > 1] lowercase = [] for duplicate_key in duplicates: lowercase = list({doc['''title'''] for doc in model_doc if doc['''local'''] == duplicate_key} ) if len(lowerCAmelCase__ ) > 1: raise ValueError( f'{duplicate_key} is present several times in the documentation table of content at ' '''`docs/source/en/_toctree.yml` with different *Title* values. Choose one of those and remove the ''' '''others.''' ) # Only add this once new_doc.append({'''local''': duplicate_key, '''title''': titles[0]} ) # Add none duplicate-keys new_doc.extend([doc for doc in model_doc if counts[doc['''local''']] == 1] ) # Sort return sorted(lowerCAmelCase__ , key=lambda lowerCAmelCase__ : s["title"].lower() ) def UpperCamelCase ( lowerCAmelCase__=False ): '''simple docstring''' with open(lowerCAmelCase__ , encoding='''utf-8''' ) as f: lowercase = yaml.safe_load(f.read() ) # Get to the API doc lowercase = 0 while content[api_idx]["title"] != "API": api_idx += 1 lowercase = content[api_idx]['''sections'''] # Then to the model doc lowercase = 0 while api_doc[model_idx]["title"] != "Models": model_idx += 1 lowercase = api_doc[model_idx]['''sections'''] lowercase = [(idx, section) for idx, section in enumerate(lowerCAmelCase__ ) if '''sections''' in section] lowercase = False for idx, modality_doc in modalities_docs: lowercase = modality_doc['''sections'''] lowercase = clean_model_doc_toc(lowerCAmelCase__ ) if old_modality_doc != new_modality_doc: lowercase = True if overwrite: lowercase = new_modality_doc if diff: if overwrite: lowercase = model_doc lowercase = api_doc with open(lowerCAmelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(yaml.dump(lowerCAmelCase__ , allow_unicode=lowerCAmelCase__ ) ) else: raise ValueError( '''The model doc part of the table of content is not properly sorted, run `make style` to fix this.''' ) if __name__ == "__main__": lowercase__ :Any = argparse.ArgumentParser() parser.add_argument("--fix_and_overwrite", action="store_true", help="Whether to fix inconsistencies.") lowercase__ :int = parser.parse_args() check_model_doc(args.fix_and_overwrite)
633
1
import json import logging import os import re import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import datasets import numpy as np import torch import torchaudio from packaging import version from torch import nn import transformers from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaProcessor, is_apex_available, set_seed, ) from transformers.trainer_utils import get_last_checkpoint, is_main_process if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse("1.6"): lowercase__ :Optional[int] = True from torch.cuda.amp import autocast lowercase__ :Any = logging.getLogger(__name__) def UpperCamelCase ( lowerCAmelCase__=None , lowerCAmelCase__=None ): '''simple docstring''' return field(default_factory=lambda: default , metadata=lowerCAmelCase__ ) @dataclass class lowercase : lowercase_ : str =field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''} ) lowercase_ : Optional[str] =field( default=SCREAMING_SNAKE_CASE__ , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) lowercase_ : Optional[bool] =field( default=SCREAMING_SNAKE_CASE__ , metadata={'''help''': '''Whether to freeze the feature extractor layers of the model.'''} ) lowercase_ : Optional[float] =field( default=0.1 , metadata={'''help''': '''The dropout ratio for the attention probabilities.'''} ) lowercase_ : Optional[float] =field( default=0.1 , metadata={'''help''': '''The dropout ratio for activations inside the fully connected layer.'''} ) lowercase_ : Optional[float] =field( default=0.1 , metadata={ '''help''': '''The dropout probabilitiy for all fully connected layers in the embeddings, encoder, and pooler.''' } , ) lowercase_ : Optional[float] =field( default=0.1 , metadata={'''help''': '''The dropout probabilitiy for all 1D convolutional layers in feature extractor.'''} , ) lowercase_ : Optional[float] =field( default=0.05 , metadata={ '''help''': ( '''Propability of each feature vector along the time axis to be chosen as the start of the vector''' '''span to be masked. Approximately ``mask_time_prob * sequence_length // mask_time_length`` feature''' '''vectors will be masked along the time axis. This is only relevant if ``apply_spec_augment is True``.''' ) } , ) lowercase_ : Optional[float] =field(default=0.0 , metadata={'''help''': '''The LayerDrop probability.'''} ) @dataclass class lowercase : lowercase_ : Optional[str] =field( default=SCREAMING_SNAKE_CASE__ , metadata={'''help''': '''The configuration name of the dataset to use (via the datasets library).'''} ) lowercase_ : Optional[str] =field( default='''train+validation''' , metadata={ '''help''': '''The name of the training data set split to use (via the datasets library). Defaults to \'train\'''' } , ) lowercase_ : bool =field( default=SCREAMING_SNAKE_CASE__ , metadata={'''help''': '''Overwrite the cached preprocessed datasets or not.'''} ) lowercase_ : Optional[int] =field( default=SCREAMING_SNAKE_CASE__ , metadata={'''help''': '''The number of processes to use for the preprocessing.'''} , ) lowercase_ : Optional[int] =field( default=SCREAMING_SNAKE_CASE__ , metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of training examples to this ''' '''value if set.''' ) } , ) lowercase_ : Optional[int] =field( default=SCREAMING_SNAKE_CASE__ , metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of validation examples to this ''' '''value if set.''' ) } , ) lowercase_ : List[str] =list_field( default=[''',''', '''?''', '''.''', '''!''', '''-''', ''';''', ''':''', '''""''', '''%''', '''\'''', '''"''', '''�'''] , metadata={'''help''': '''A list of characters to remove from the transcripts.'''} , ) @dataclass class lowercase : lowercase_ : WavaVecaProcessor lowercase_ : Union[bool, str] =True lowercase_ : Optional[int] =None lowercase_ : Optional[int] =None lowercase_ : Optional[int] =None lowercase_ : Optional[int] =None def __call__( self ,A__): # split inputs and labels since they have to be of different lenghts and need # different padding methods lowercase = [{'''input_values''': feature['''input_values''']} for feature in features] lowercase = [{'''input_ids''': feature['''labels''']} for feature in features] lowercase = self.processor.pad( A__ ,padding=self.padding ,max_length=self.max_length ,pad_to_multiple_of=self.pad_to_multiple_of ,return_tensors='''pt''' ,) lowercase = self.processor.pad( labels=A__ ,padding=self.padding ,max_length=self.max_length_labels ,pad_to_multiple_of=self.pad_to_multiple_of_labels ,return_tensors='''pt''' ,) # replace padding with -100 to ignore loss correctly lowercase = labels_batch['''input_ids'''].masked_fill(labels_batch.attention_mask.ne(1) ,-1_0_0) lowercase = labels return batch class lowercase ( SCREAMING_SNAKE_CASE__ ): def A__ ( self ,A__ ,A__): model.train() lowercase = self._prepare_inputs(A__) if self.use_amp: with autocast(): lowercase = self.compute_loss(A__ ,A__) else: lowercase = self.compute_loss(A__ ,A__) if self.args.n_gpu > 1: if model.module.config.ctc_loss_reduction == "mean": lowercase = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": lowercase = loss.sum() / (inputs['''labels'''] >= 0).sum() else: raise ValueError(f'{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']') if self.args.gradient_accumulation_steps > 1: lowercase = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(A__).backward() elif self.use_apex: with amp.scale_loss(A__ ,self.optimizer) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(A__) else: loss.backward() return loss.detach() def UpperCamelCase ( ): '''simple docstring''' # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowercase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. lowercase , lowercase , lowercase = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: lowercase , lowercase , lowercase = parser.parse_args_into_dataclasses() # Detecting last checkpoint. lowercase = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: lowercase = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f'Output directory ({training_args.output_dir}) already exists and is not empty. ' '''Use --overwrite_output_dir to overcome.''' ) elif last_checkpoint is not None: logger.info( f'Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ' '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) logger.setLevel(logging.INFO if is_main_process(training_args.local_rank ) else logging.WARN ) # Log on each process the small summary: logger.warning( f'Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}' + f'distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}' ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() logger.info('''Training/evaluation parameters %s''' , lowerCAmelCase__ ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: lowercase = datasets.load_dataset( '''common_voice''' , data_args.dataset_config_name , split=data_args.train_split_name ) lowercase = datasets.load_dataset('''common_voice''' , data_args.dataset_config_name , split='''test''' ) # Create and save tokenizer lowercase = f'[{"".join(data_args.chars_to_ignore )}]' def remove_special_characters(lowerCAmelCase__ ): lowercase = re.sub(lowerCAmelCase__ , '''''' , batch['''sentence'''] ).lower() + ''' ''' return batch lowercase = train_dataset.map(lowerCAmelCase__ , remove_columns=['''sentence'''] ) lowercase = eval_dataset.map(lowerCAmelCase__ , remove_columns=['''sentence'''] ) def extract_all_chars(lowerCAmelCase__ ): lowercase = ''' '''.join(batch['''text'''] ) lowercase = list(set(lowerCAmelCase__ ) ) return {"vocab": [vocab], "all_text": [all_text]} lowercase = train_dataset.map( lowerCAmelCase__ , batched=lowerCAmelCase__ , batch_size=-1 , keep_in_memory=lowerCAmelCase__ , remove_columns=train_dataset.column_names , ) lowercase = train_dataset.map( lowerCAmelCase__ , batched=lowerCAmelCase__ , batch_size=-1 , keep_in_memory=lowerCAmelCase__ , remove_columns=eval_dataset.column_names , ) lowercase = list(set(vocab_train['''vocab'''][0] ) | set(vocab_test['''vocab'''][0] ) ) lowercase = {v: k for k, v in enumerate(lowerCAmelCase__ )} lowercase = vocab_dict[''' '''] del vocab_dict[" "] lowercase = len(lowerCAmelCase__ ) lowercase = len(lowerCAmelCase__ ) with open('''vocab.json''' , '''w''' ) as vocab_file: json.dump(lowerCAmelCase__ , lowerCAmelCase__ ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowercase = WavaVecaCTCTokenizer( '''vocab.json''' , unk_token='''[UNK]''' , pad_token='''[PAD]''' , word_delimiter_token='''|''' , ) lowercase = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6000 , padding_value=0.0 , do_normalize=lowerCAmelCase__ , return_attention_mask=lowerCAmelCase__ ) lowercase = WavaVecaProcessor(feature_extractor=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ ) lowercase = WavaVecaForCTC.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , activation_dropout=model_args.activation_dropout , attention_dropout=model_args.attention_dropout , hidden_dropout=model_args.hidden_dropout , feat_proj_dropout=model_args.feat_proj_dropout , mask_time_prob=model_args.mask_time_prob , gradient_checkpointing=training_args.gradient_checkpointing , layerdrop=model_args.layerdrop , ctc_loss_reduction='''mean''' , pad_token_id=processor.tokenizer.pad_token_id , vocab_size=len(processor.tokenizer ) , ) if data_args.max_train_samples is not None: lowercase = min(len(lowerCAmelCase__ ) , data_args.max_train_samples ) lowercase = train_dataset.select(range(lowerCAmelCase__ ) ) if data_args.max_val_samples is not None: lowercase = eval_dataset.select(range(data_args.max_val_samples ) ) lowercase = torchaudio.transforms.Resample(4_8000 , 1_6000 ) # Preprocessing the datasets. # We need to read the aduio files as arrays and tokenize the targets. def speech_file_to_array_fn(lowerCAmelCase__ ): lowercase , lowercase = torchaudio.load(batch['''path'''] ) lowercase = resampler(lowerCAmelCase__ ).squeeze().numpy() lowercase = 1_6000 lowercase = batch['''text'''] return batch lowercase = train_dataset.map( lowerCAmelCase__ , remove_columns=train_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) lowercase = eval_dataset.map( lowerCAmelCase__ , remove_columns=eval_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) def prepare_dataset(lowerCAmelCase__ ): # check that all files have the correct sampling rate assert ( len(set(batch['''sampling_rate'''] ) ) == 1 ), f'Make sure all inputs have the same sampling rate of {processor.feature_extractor.sampling_rate}.' lowercase = processor( audio=batch['''speech'''] , text=batch['''target_text'''] , sampling_rate=batch['''sampling_rate'''][0] ) batch.update(lowerCAmelCase__ ) return batch lowercase = train_dataset.map( lowerCAmelCase__ , remove_columns=train_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=lowerCAmelCase__ , num_proc=data_args.preprocessing_num_workers , ) lowercase = eval_dataset.map( lowerCAmelCase__ , remove_columns=eval_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=lowerCAmelCase__ , num_proc=data_args.preprocessing_num_workers , ) # Metric lowercase = datasets.load_metric('''wer''' ) def compute_metrics(lowerCAmelCase__ ): lowercase = pred.predictions lowercase = np.argmax(lowerCAmelCase__ , axis=-1 ) lowercase = processor.tokenizer.pad_token_id lowercase = processor.batch_decode(lowerCAmelCase__ ) # we do not want to group tokens when computing the metrics lowercase = processor.batch_decode(pred.label_ids , group_tokens=lowerCAmelCase__ ) lowercase = wer_metric.compute(predictions=lowerCAmelCase__ , references=lowerCAmelCase__ ) return {"wer": wer} if model_args.freeze_feature_extractor: model.freeze_feature_extractor() # Data collator lowercase = DataCollatorCTCWithPadding(processor=lowerCAmelCase__ , padding=lowerCAmelCase__ ) # Initialize our Trainer lowercase = CTCTrainer( model=lowerCAmelCase__ , data_collator=lowerCAmelCase__ , args=lowerCAmelCase__ , compute_metrics=lowerCAmelCase__ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=processor.feature_extractor , ) # Training if training_args.do_train: if last_checkpoint is not None: lowercase = last_checkpoint elif os.path.isdir(model_args.model_name_or_path ): lowercase = model_args.model_name_or_path else: lowercase = None # Save the feature_extractor and the tokenizer if is_main_process(training_args.local_rank ): processor.save_pretrained(training_args.output_dir ) lowercase = trainer.train(resume_from_checkpoint=lowerCAmelCase__ ) trainer.save_model() lowercase = train_result.metrics lowercase = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(lowerCAmelCase__ ) ) lowercase = min(lowerCAmelCase__ , len(lowerCAmelCase__ ) ) trainer.log_metrics('''train''' , lowerCAmelCase__ ) trainer.save_metrics('''train''' , lowerCAmelCase__ ) trainer.save_state() # Evaluation lowercase = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) lowercase = trainer.evaluate() lowercase = data_args.max_val_samples if data_args.max_val_samples is not None else len(lowerCAmelCase__ ) lowercase = min(lowerCAmelCase__ , len(lowerCAmelCase__ ) ) trainer.log_metrics('''eval''' , lowerCAmelCase__ ) trainer.save_metrics('''eval''' , lowerCAmelCase__ ) return results if __name__ == "__main__": main()
633
from __future__ import annotations from decimal import Decimal from math import * # noqa: F403 from sympy import diff def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 10**-10 ): '''simple docstring''' lowercase = a while True: lowercase = Decimal(lowerCAmelCase__ ) - ( Decimal(eval(lowerCAmelCase__ ) ) / Decimal(eval(str(diff(lowerCAmelCase__ ) ) ) ) # noqa: S307 ) # This number dictates the accuracy of the answer if abs(eval(lowerCAmelCase__ ) ) < precision: # noqa: S307 return float(lowerCAmelCase__ ) # Let's Execute if __name__ == "__main__": # Find root of trigonometric function # Find value of pi print(F'The root of sin(x) = 0 is {newton_raphson("sin(x)", 2)}') # Find root of polynomial print(F'The root of x**2 - 5*x + 2 = 0 is {newton_raphson("x**2 - 5*x + 2", 0.4)}') # Find Square Root of 5 print(F'The root of log(x) - 1 = 0 is {newton_raphson("log(x) - 1", 2)}') # Exponential Roots print(F'The root of exp(x) - 1 = 0 is {newton_raphson("exp(x) - 1", 0)}')
633
1
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class lowercase ( unittest.TestCase ): @slow def A__ ( self): lowercase = TFCamembertModel.from_pretrained('''jplu/tf-camembert-base''') lowercase = tf.convert_to_tensor( [[5, 1_2_1, 1_1, 6_6_0, 1_6, 7_3_0, 2_5_5_4_3, 1_1_0, 8_3, 6]] ,dtype=tf.intaa ,) # J'aime le camembert !" lowercase = model(A__)['''last_hidden_state'''] lowercase = tf.TensorShape((1, 1_0, 7_6_8)) self.assertEqual(output.shape ,A__) # compare the actual values for a slice. lowercase = tf.convert_to_tensor( [[[-0.0254, 0.0235, 0.1027], [0.0606, -0.1811, -0.0418], [-0.1561, -0.1127, 0.2687]]] ,dtype=tf.floataa ,) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() ,expected_slice.numpy() ,atol=1E-4))
633
from pathlib import Path import numpy as np from PIL import Image def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase , lowercase = rgb[:, :, 0], rgb[:, :, 1], rgb[:, :, 2] return 0.29_89 * r + 0.58_70 * g + 0.11_40 * b def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return (gray > 127) & (gray <= 255) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = np.zeros_like(lowerCAmelCase__ ) lowercase = np.zeros( (image.shape[0] + kernel.shape[0] - 1, image.shape[1] + kernel.shape[1] - 1) ) # Copy image to padded image lowercase = image # Iterate over image & apply kernel for x in range(image.shape[1] ): for y in range(image.shape[0] ): lowercase = ( kernel * image_padded[y : y + kernel.shape[0], x : x + kernel.shape[1]] ).sum() lowercase = int(summation > 0 ) return output if __name__ == "__main__": # read original image lowercase__ :str = Path(__file__).resolve().parent / "image_data" / "lena.jpg" lowercase__ :List[str] = np.array(Image.open(lena_path)) # kernel to be applied lowercase__ :Union[str, Any] = np.array([[0, 1, 0], [1, 1, 1], [0, 1, 0]]) lowercase__ :Optional[int] = dilation(gray_to_binary(rgb_to_gray(lena)), structuring_element) # Save the output image lowercase__ :str = Image.fromarray(output).convert("RGB") pil_img.save("result_dilation.png")
633
1
import inspect import os import unittest import torch import accelerate from accelerate import debug_launcher from accelerate.test_utils import ( execute_subprocess_async, require_cpu, require_huggingface_suite, require_multi_gpu, require_single_gpu, ) from accelerate.utils import patch_environment @require_huggingface_suite class lowercase ( unittest.TestCase ): def A__ ( self): lowercase = inspect.getfile(accelerate.test_utils) lowercase = os.path.sep.join( mod_file.split(os.path.sep)[:-1] + ['''scripts''', '''external_deps''', '''test_metrics.py''']) from accelerate.test_utils.scripts.external_deps import test_metrics # noqa: F401 lowercase = test_metrics @require_cpu def A__ ( self): debug_launcher(self.test_metrics.main ,num_processes=1) @require_cpu def A__ ( self): debug_launcher(self.test_metrics.main) @require_single_gpu def A__ ( self): self.test_metrics.main() @require_multi_gpu def A__ ( self): print(f'Found {torch.cuda.device_count()} devices.') lowercase = ['''torchrun''', f'--nproc_per_node={torch.cuda.device_count()}', self.test_file_path] with patch_environment(omp_num_threads=1): execute_subprocess_async(A__ ,env=os.environ.copy())
633
import argparse import torch from torch import nn from transformers import MaMaaaConfig, MaMaaaForConditionalGeneration def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = [ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''decoder.output_projection.weight''', '''_float_tensor''', '''encoder.embed_positions._float_tensor''', '''decoder.embed_positions._float_tensor''', ] for k in ignore_keys: state_dict.pop(lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase = emb.weight.shape lowercase = nn.Linear(lowerCAmelCase__ , lowerCAmelCase__ , bias=lowerCAmelCase__ ) lowercase = emb.weight.data return lin_layer def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = torch.load(lowerCAmelCase__ , map_location='''cpu''' ) lowercase = mam_aaa['''args'''] or mam_aaa['''cfg''']['''model'''] lowercase = mam_aaa['''model'''] remove_ignore_keys_(lowerCAmelCase__ ) lowercase = state_dict['''encoder.embed_tokens.weight'''].shape[0] lowercase = MaMaaaConfig( vocab_size=lowerCAmelCase__ , max_position_embeddings=1024 , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , encoder_layerdrop=args.encoder_layerdrop , decoder_layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='''relu''' , ) lowercase = state_dict['''decoder.embed_tokens.weight'''] lowercase = MaMaaaForConditionalGeneration(lowerCAmelCase__ ) model.model.load_state_dict(lowerCAmelCase__ , strict=lowerCAmelCase__ ) lowercase = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": lowercase__ :Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument("fairseq_path", type=str, help="path to a model.pt on local filesystem.") parser.add_argument("pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") lowercase__ :Tuple = parser.parse_args() lowercase__ :int = convert_fairseq_mamaaa_checkpoint_from_disk(args.fairseq_pathß) model.save_pretrained(args.pytorch_dump_folder_path)
633
1
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFXLMRobertaModel @require_tf @require_sentencepiece @require_tokenizers class lowercase ( unittest.TestCase ): @slow def A__ ( self): lowercase = TFXLMRobertaModel.from_pretrained('''jplu/tf-xlm-roberta-base''') lowercase = { '''input_ids''': tf.convert_to_tensor([[0, 2_6_4_6, 1_0_2_6_9, 8_3, 9_9_9_4_2, 2]] ,dtype=tf.intaa), # "My dog is cute" '''attention_mask''': tf.convert_to_tensor([[1, 1, 1, 1, 1, 1]] ,dtype=tf.intaa), } lowercase = model(A__)['''last_hidden_state'''] lowercase = tf.TensorShape((1, 6, 7_6_8)) self.assertEqual(output.shape ,A__) # compare the actual values for a slice. lowercase = tf.convert_to_tensor( [ [ [0.0681762, 0.10894451, 0.06772504], [-0.06423668, 0.02366615, 0.04329344], [-0.06057295, 0.09974135, -0.00070584], ] ] ,dtype=tf.floataa ,) self.assertTrue(np.allclose(output[:, :3, :3].numpy() ,expected_slice.numpy() ,atol=1E-4))
633
from __future__ import annotations from random import random class lowercase : def __init__( self ,A__ = None): lowercase = value lowercase = random() lowercase = None lowercase = None def __repr__( self): from pprint import pformat if self.left is None and self.right is None: return f'\'{self.value}: {self.prior:.5}\'' else: return pformat( {f'{self.value}: {self.prior:.5}': (self.left, self.right)} ,indent=1) def __str__( self): lowercase = str(self.value) + ''' ''' lowercase = str(self.left or '''''') lowercase = str(self.right or '''''') return value + left + right def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' if root is None: # None tree is split into 2 Nones return None, None elif root.value is None: return None, None else: if value < root.value: lowercase , lowercase = split(root.left , lowerCAmelCase__ ) return left, root else: lowercase , lowercase = split(root.right , lowerCAmelCase__ ) return root, right def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' if (not left) or (not right): # If one node is None, return the other return left or right elif left.prior < right.prior: lowercase = merge(left.right , lowerCAmelCase__ ) return left else: lowercase = merge(lowerCAmelCase__ , right.left ) return right def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = Node(lowerCAmelCase__ ) lowercase , lowercase = split(lowerCAmelCase__ , lowerCAmelCase__ ) return merge(merge(lowerCAmelCase__ , lowerCAmelCase__ ) , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase = split(lowerCAmelCase__ , value - 1 ) lowercase , lowercase = split(lowerCAmelCase__ , lowerCAmelCase__ ) return merge(lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if not root: # None return else: inorder(root.left ) print(root.value , end=''',''' ) inorder(root.right ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' for arg in args.split(): if arg[0] == "+": lowercase = insert(lowerCAmelCase__ , int(arg[1:] ) ) elif arg[0] == "-": lowercase = erase(lowerCAmelCase__ , int(arg[1:] ) ) else: print('''Unknown command''' ) return root def UpperCamelCase ( ): '''simple docstring''' lowercase = None print( '''enter numbers to create a tree, + value to add value into treap, ''' '''- value to erase all nodes with value. \'q\' to quit. ''' ) lowercase = input() while args != "q": lowercase = interact_treap(lowerCAmelCase__ , lowerCAmelCase__ ) print(lowerCAmelCase__ ) lowercase = input() print('''good by!''' ) if __name__ == "__main__": import doctest doctest.testmod() main()
633
1
import html from ...feature_extraction_utils import BatchFeature, FeatureExtractionMixin from ...utils import is_bsa_available, logging, requires_backends if is_bsa_available(): import bsa from bsa import BeautifulSoup lowercase__ :str = logging.get_logger(__name__) class lowercase ( SCREAMING_SNAKE_CASE__ ): def __init__( self ,**A__): requires_backends(self ,['''bs4''']) super().__init__(**A__) def A__ ( self ,A__): lowercase = [] lowercase = [] lowercase = element if element.name else element.parent for parent in child.parents: # type: bs4.element.Tag lowercase = parent.find_all(child.name ,recursive=A__) xpath_tags.append(child.name) xpath_subscripts.append( 0 if 1 == len(A__) else next(i for i, s in enumerate(A__ ,1) if s is child)) lowercase = parent xpath_tags.reverse() xpath_subscripts.reverse() return xpath_tags, xpath_subscripts def A__ ( self ,A__): lowercase = BeautifulSoup(A__ ,'''html.parser''') lowercase = [] lowercase = [] lowercase = [] for element in html_code.descendants: if type(A__) == bsa.element.NavigableString: if type(element.parent) != bsa.element.Tag: continue lowercase = html.unescape(A__).strip() if not text_in_this_tag: continue all_doc_strings.append(A__) lowercase , lowercase = self.xpath_soup(A__) stringaxtag_seq.append(A__) stringaxsubs_seq.append(A__) if len(A__) != len(A__): raise ValueError('''Number of doc strings and xtags does not correspond''') if len(A__) != len(A__): raise ValueError('''Number of doc strings and xsubs does not correspond''') return all_doc_strings, stringaxtag_seq, stringaxsubs_seq def A__ ( self ,A__ ,A__): lowercase = '''''' for tagname, subs in zip(A__ ,A__): xpath += f'/{tagname}' if subs != 0: xpath += f'[{subs}]' return xpath def __call__( self ,A__): lowercase = False # Check that strings has a valid type if isinstance(A__ ,A__): lowercase = True elif isinstance(A__ ,(list, tuple)): if len(A__) == 0 or isinstance(html_strings[0] ,A__): lowercase = True if not valid_strings: raise ValueError( '''HTML strings must of type `str`, `List[str]` (batch of examples), ''' f'but is of type {type(A__)}.') lowercase = bool(isinstance(A__ ,(list, tuple)) and (isinstance(html_strings[0] ,A__))) if not is_batched: lowercase = [html_strings] # Get nodes + xpaths lowercase = [] lowercase = [] for html_string in html_strings: lowercase , lowercase , lowercase = self.get_three_from_single(A__) nodes.append(A__) lowercase = [] for node, tag_list, sub_list in zip(A__ ,A__ ,A__): lowercase = self.construct_xpath(A__ ,A__) xpath_strings.append(A__) xpaths.append(A__) # return as Dict lowercase = {'''nodes''': nodes, '''xpaths''': xpaths} lowercase = BatchFeature(data=A__ ,tensor_type=A__) return encoded_inputs
633
def UpperCamelCase ( lowerCAmelCase__ = 1000 ): '''simple docstring''' lowercase = -1 lowercase = 0 for a in range(1 , n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N eliminating c lowercase = (n * n - 2 * a * n) // (2 * n - 2 * a) lowercase = n - a - b if c * c == (a * a + b * b): lowercase = a * b * c if candidate >= product: lowercase = candidate return product if __name__ == "__main__": print(F'{solution() = }')
633
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase__ :Any = logging.get_logger(__name__) lowercase__ :str = { "google/vit-base-patch16-224": "https://huggingface.co/vit-base-patch16-224/resolve/main/config.json", # See all ViT models at https://huggingface.co/models?filter=vit } class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : List[Any] ='''vit''' def __init__( self ,A__=7_6_8 ,A__=1_2 ,A__=1_2 ,A__=3_0_7_2 ,A__="gelu" ,A__=0.0 ,A__=0.0 ,A__=0.02 ,A__=1E-12 ,A__=2_2_4 ,A__=1_6 ,A__=3 ,A__=True ,A__=1_6 ,**A__ ,): super().__init__(**A__) lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_attention_heads lowercase = intermediate_size lowercase = hidden_act lowercase = hidden_dropout_prob lowercase = attention_probs_dropout_prob lowercase = initializer_range lowercase = layer_norm_eps lowercase = image_size lowercase = patch_size lowercase = num_channels lowercase = qkv_bias lowercase = encoder_stride class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Optional[int] =version.parse('''1.11''' ) @property def A__ ( self): return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ]) @property def A__ ( self): return 1E-4
633
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ :Tuple = { "configuration_instructblip": [ "INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "InstructBlipConfig", "InstructBlipQFormerConfig", "InstructBlipVisionConfig", ], "processing_instructblip": ["InstructBlipProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :List[str] = [ "INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "InstructBlipQFormerModel", "InstructBlipPreTrainedModel", "InstructBlipForConditionalGeneration", "InstructBlipVisionModel", ] if TYPE_CHECKING: from .configuration_instructblip import ( INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, InstructBlipConfig, InstructBlipQFormerConfig, InstructBlipVisionConfig, ) from .processing_instructblip import InstructBlipProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_instructblip import ( INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST, InstructBlipForConditionalGeneration, InstructBlipPreTrainedModel, InstructBlipQFormerModel, InstructBlipVisionModel, ) else: import sys lowercase__ :List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
633
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_lxmert import LxmertTokenizer lowercase__ :str = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} lowercase__ :List[Any] = { "vocab_file": { "unc-nlp/lxmert-base-uncased": "https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/vocab.txt", }, "tokenizer_file": { "unc-nlp/lxmert-base-uncased": ( "https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/tokenizer.json" ), }, } lowercase__ :Union[str, Any] = { "unc-nlp/lxmert-base-uncased": 512, } lowercase__ :str = { "unc-nlp/lxmert-base-uncased": {"do_lower_case": True}, } class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Tuple =VOCAB_FILES_NAMES lowercase_ : int =PRETRAINED_VOCAB_FILES_MAP lowercase_ : Any =PRETRAINED_INIT_CONFIGURATION lowercase_ : Any =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ : Dict =LxmertTokenizer def __init__( self ,A__=None ,A__=None ,A__=True ,A__="[UNK]" ,A__="[SEP]" ,A__="[PAD]" ,A__="[CLS]" ,A__="[MASK]" ,A__=True ,A__=None ,**A__ ,): super().__init__( A__ ,tokenizer_file=A__ ,do_lower_case=A__ ,unk_token=A__ ,sep_token=A__ ,pad_token=A__ ,cls_token=A__ ,mask_token=A__ ,tokenize_chinese_chars=A__ ,strip_accents=A__ ,**A__ ,) lowercase = json.loads(self.backend_tokenizer.normalizer.__getstate__()) if ( normalizer_state.get('''lowercase''' ,A__) != do_lower_case or normalizer_state.get('''strip_accents''' ,A__) != strip_accents or normalizer_state.get('''handle_chinese_chars''' ,A__) != tokenize_chinese_chars ): lowercase = getattr(A__ ,normalizer_state.pop('''type''')) lowercase = do_lower_case lowercase = strip_accents lowercase = tokenize_chinese_chars lowercase = normalizer_class(**A__) lowercase = do_lower_case def A__ ( self ,A__ ,A__=None): lowercase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A__ ( self ,A__ ,A__ = None): lowercase = [self.sep_token_id] lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def A__ ( self ,A__ ,A__ = None): lowercase = self._tokenizer.model.save(A__ ,name=A__) return tuple(A__)
633
from numpy import exp, pi, sqrt def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ = 0.0 , lowerCAmelCase__ = 1.0 ): '''simple docstring''' return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
633
1
from graphs.minimum_spanning_tree_kruskal import kruskal def UpperCamelCase ( ): '''simple docstring''' lowercase = 9 lowercase = [ [0, 1, 4], [0, 7, 8], [1, 2, 8], [7, 8, 7], [7, 6, 1], [2, 8, 2], [8, 6, 6], [2, 3, 7], [2, 5, 4], [6, 5, 2], [3, 5, 14], [3, 4, 9], [5, 4, 10], [1, 7, 11], ] lowercase = kruskal(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase = [ [7, 6, 1], [2, 8, 2], [6, 5, 2], [0, 1, 4], [2, 5, 4], [2, 3, 7], [0, 7, 8], [3, 4, 9], ] assert sorted(lowerCAmelCase__ ) == sorted(lowerCAmelCase__ )
633
import asyncio import os import shutil import subprocess import sys import tempfile import unittest from distutils.util import strtobool from functools import partial from pathlib import Path from typing import List, Union from unittest import mock import torch from ..state import AcceleratorState, PartialState from ..utils import ( gather, is_bnb_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_mps_available, is_safetensors_available, is_tensorboard_available, is_torch_version, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=False ): '''simple docstring''' try: lowercase = os.environ[key] except KeyError: # KEY isn't set, default to `default`. lowercase = default else: # KEY is set, convert it to True or False. try: lowercase = strtobool(lowerCAmelCase__ ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(f'If set, {key} must be yes or no.' ) return _value lowercase__ :Dict = parse_flag_from_env("RUN_SLOW", default=False) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skip('''Test was skipped''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(_run_slow_tests , '''test is slow''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(not torch.cuda.is_available() , '''test requires only a CPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.cuda.is_available() , '''test requires a GPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_xpu_available() , '''test requires a XPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_mps_available() , '''test requires a `mps` backend support in `torch`''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless( is_transformers_available() and is_datasets_available() , '''test requires the Hugging Face suite''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_bnb_available() , '''test requires the bitsandbytes library''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_tpu_available() , '''test requires TPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.cuda.device_count() == 1 , '''test requires a GPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.xpu.device_count() == 1 , '''test requires a XPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.cuda.device_count() > 1 , '''test requires multiple GPUs''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.xpu.device_count() > 1 , '''test requires multiple XPUs''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_safetensors_available() , '''test requires safetensors''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_deepspeed_available() , '''test requires DeepSpeed''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_torch_version('''>=''' , '''1.12.0''' ) , '''test requires torch version >= 1.12.0''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__=None , lowerCAmelCase__=None ): '''simple docstring''' if test_case is None: return partial(lowerCAmelCase__ , version=lowerCAmelCase__ ) return unittest.skipUnless(is_torch_version('''>=''' , lowerCAmelCase__ ) , f'test requires torch version >= {version}' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_tensorboard_available() , '''test requires Tensorboard''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_wandb_available() , '''test requires wandb''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_comet_ml_available() , '''test requires comet_ml''' )(lowerCAmelCase__ ) lowercase__ :Dict = ( any([is_wandb_available(), is_tensorboard_available()]) and not is_comet_ml_available() ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless( _atleast_one_tracker_available , '''test requires at least one tracker to be available and for `comet_ml` to not be installed''' , )(lowerCAmelCase__ ) class lowercase ( unittest.TestCase ): lowercase_ : int =True @classmethod def A__ ( cls): lowercase = tempfile.mkdtemp() @classmethod def A__ ( cls): if os.path.exists(cls.tmpdir): shutil.rmtree(cls.tmpdir) def A__ ( self): if self.clear_on_setup: for path in Path(self.tmpdir).glob('''**/*'''): if path.is_file(): path.unlink() elif path.is_dir(): shutil.rmtree(A__) class lowercase ( unittest.TestCase ): def A__ ( self): super().tearDown() # Reset the state of the AcceleratorState singleton. AcceleratorState._reset_state() PartialState._reset_state() class lowercase ( unittest.TestCase ): def A__ ( self ,A__): lowercase = mocks if isinstance(A__ ,(tuple, list)) else [mocks] for m in self.mocks: m.start() self.addCleanup(m.stop) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = AcceleratorState() lowercase = tensor[None].clone().to(state.device ) lowercase = gather(lowerCAmelCase__ ).cpu() lowercase = tensor[0].cpu() for i in range(tensors.shape[0] ): if not torch.equal(tensors[i] , lowerCAmelCase__ ): return False return True class lowercase : def __init__( self ,A__ ,A__ ,A__): lowercase = returncode lowercase = stdout lowercase = stderr async def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' while True: lowercase = await stream.readline() if line: callback(lowerCAmelCase__ ) else: break async def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=False , lowerCAmelCase__=False ): '''simple docstring''' if echo: print('''\nRunning: ''' , ''' '''.join(lowerCAmelCase__ ) ) lowercase = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=lowerCAmelCase__ , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=lowerCAmelCase__ , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) lowercase = [] lowercase = [] def tee(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__="" ): lowercase = line.decode('''utf-8''' ).rstrip() sink.append(lowerCAmelCase__ ) if not quiet: print(lowerCAmelCase__ , lowerCAmelCase__ , file=lowerCAmelCase__ ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ asyncio.create_task(_read_stream(p.stdout , lambda lowerCAmelCase__ : tee(lowerCAmelCase__ , lowerCAmelCase__ , sys.stdout , label='''stdout:''' ) ) ), asyncio.create_task(_read_stream(p.stderr , lambda lowerCAmelCase__ : tee(lowerCAmelCase__ , lowerCAmelCase__ , sys.stderr , label='''stderr:''' ) ) ), ] , timeout=lowerCAmelCase__ , ) return _RunOutput(await p.wait() , lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=180 , lowerCAmelCase__=False , lowerCAmelCase__=True ): '''simple docstring''' lowercase = asyncio.get_event_loop() lowercase = loop.run_until_complete( _stream_subprocess(lowerCAmelCase__ , env=lowerCAmelCase__ , stdin=lowerCAmelCase__ , timeout=lowerCAmelCase__ , quiet=lowerCAmelCase__ , echo=lowerCAmelCase__ ) ) lowercase = ''' '''.join(lowerCAmelCase__ ) if result.returncode > 0: lowercase = '''\n'''.join(result.stderr ) raise RuntimeError( f'\'{cmd_str}\' failed with returncode {result.returncode}\n\n' f'The combined stderr from workers follows:\n{stderr}' ) return result class lowercase ( SCREAMING_SNAKE_CASE__ ): pass def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=False ): '''simple docstring''' try: lowercase = subprocess.check_output(lowerCAmelCase__ , stderr=subprocess.STDOUT ) if return_stdout: if hasattr(lowerCAmelCase__ , '''decode''' ): lowercase = output.decode('''utf-8''' ) return output except subprocess.CalledProcessError as e: raise SubprocessCallException( f'Command `{" ".join(lowerCAmelCase__ )}` failed with the following error:\n\n{e.output.decode()}' ) from e
633
1
from collections.abc import Iterable from typing import Generic, TypeVar lowercase__ :Any = TypeVar("_T") class lowercase ( Generic[_T] ): def __init__( self ,A__ = None): lowercase = list(iterable or []) lowercase = [] def __len__( self): return len(self._stacka) + len(self._stacka) def __repr__( self): return f'Queue({tuple(self._stacka[::-1] + self._stacka)})' def A__ ( self ,A__): self._stacka.append(A__) def A__ ( self): lowercase = self._stacka.pop lowercase = self._stacka.append if not self._stacka: while self._stacka: stacka_append(stacka_pop()) if not self._stacka: raise IndexError('''Queue is empty''') return self._stacka.pop() if __name__ == "__main__": from doctest import testmod testmod()
633
import math import time from typing import Dict, List, Optional from torch.utils.data import Dataset from transformers import SeqaSeqTrainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class lowercase ( SCREAMING_SNAKE_CASE__ ): def __init__( self ,*A__ ,A__=None ,A__=None ,**A__): super().__init__(*A__ ,**A__) lowercase = eval_examples lowercase = post_process_function def A__ ( self ,A__ = None ,A__=None ,A__ = None ,A__ = "eval" ,**A__ ,): lowercase = gen_kwargs.copy() lowercase = ( gen_kwargs['''max_length'''] if gen_kwargs.get('''max_length''') is not None else self.args.generation_max_length ) lowercase = ( gen_kwargs['''num_beams'''] if gen_kwargs.get('''num_beams''') is not None else self.args.generation_num_beams ) lowercase = gen_kwargs lowercase = self.eval_dataset if eval_dataset is None else eval_dataset lowercase = self.get_eval_dataloader(A__) lowercase = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. lowercase = self.compute_metrics lowercase = None lowercase = time.time() lowercase = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: lowercase = eval_loop( A__ ,description='''Evaluation''' ,prediction_loss_only=True if compute_metrics is None else None ,ignore_keys=A__ ,metric_key_prefix=A__ ,) finally: lowercase = compute_metrics lowercase = self.args.eval_batch_size * self.args.world_size if f'{metric_key_prefix}_jit_compilation_time' in output.metrics: start_time += output.metrics[f'{metric_key_prefix}_jit_compilation_time'] output.metrics.update( speed_metrics( A__ ,A__ ,num_samples=output.num_samples ,num_steps=math.ceil(output.num_samples / total_batch_size) ,)) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default lowercase = self.post_process_function(A__ ,A__ ,A__) lowercase = self.compute_metrics(A__) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys()): if not key.startswith(f'{metric_key_prefix}_'): lowercase = metrics.pop(A__) metrics.update(output.metrics) else: lowercase = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(A__) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report()) lowercase = self.callback_handler.on_evaluate(self.args ,self.state ,self.control ,A__) return metrics def A__ ( self ,A__ ,A__ ,A__=None ,A__ = "test" ,**A__): lowercase = gen_kwargs.copy() lowercase = self.get_test_dataloader(A__) # Temporarily disable metric computation, we will do it in the loop here. lowercase = self.compute_metrics lowercase = None lowercase = time.time() lowercase = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: lowercase = eval_loop( A__ ,description='''Prediction''' ,prediction_loss_only=True if compute_metrics is None else None ,ignore_keys=A__ ,metric_key_prefix=A__ ,) finally: lowercase = compute_metrics lowercase = self.args.eval_batch_size * self.args.world_size if f'{metric_key_prefix}_jit_compilation_time' in output.metrics: start_time += output.metrics[f'{metric_key_prefix}_jit_compilation_time'] output.metrics.update( speed_metrics( A__ ,A__ ,num_samples=output.num_samples ,num_steps=math.ceil(output.num_samples / total_batch_size) ,)) if self.post_process_function is None or self.compute_metrics is None: return output lowercase = self.post_process_function(A__ ,A__ ,A__ ,'''predict''') lowercase = self.compute_metrics(A__) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys()): if not key.startswith(f'{metric_key_prefix}_'): lowercase = metrics.pop(A__) metrics.update(output.metrics) return PredictionOutput(predictions=predictions.predictions ,label_ids=predictions.label_ids ,metrics=A__)
633
1
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ :List[Any] = logging.get_logger(__name__) lowercase__ :List[Any] = { "microsoft/unispeech-sat-base-100h-libri-ft": ( "https://huggingface.co/microsoft/unispeech-sat-base-100h-libri-ft/resolve/main/config.json" ), # See all UniSpeechSat models at https://huggingface.co/models?filter=unispeech_sat } class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Union[str, Any] ='''unispeech-sat''' def __init__( self ,A__=3_2 ,A__=7_6_8 ,A__=1_2 ,A__=1_2 ,A__=3_0_7_2 ,A__="gelu" ,A__=0.1 ,A__=0.1 ,A__=0.1 ,A__=0.0 ,A__=0.0 ,A__=0.1 ,A__=0.1 ,A__=0.02 ,A__=1E-5 ,A__="group" ,A__="gelu" ,A__=(5_1_2, 5_1_2, 5_1_2, 5_1_2, 5_1_2, 5_1_2, 5_1_2) ,A__=(5, 2, 2, 2, 2, 2, 2) ,A__=(1_0, 3, 3, 3, 3, 2, 2) ,A__=False ,A__=1_2_8 ,A__=1_6 ,A__=False ,A__=True ,A__=0.05 ,A__=1_0 ,A__=2 ,A__=0.0 ,A__=1_0 ,A__=0 ,A__=3_2_0 ,A__=2 ,A__=0.1 ,A__=1_0_0 ,A__=2_5_6 ,A__=2_5_6 ,A__=0.1 ,A__="mean" ,A__=False ,A__=False ,A__=2_5_6 ,A__=(5_1_2, 5_1_2, 5_1_2, 5_1_2, 1_5_0_0) ,A__=(5, 3, 3, 1, 1) ,A__=(1, 2, 3, 1, 1) ,A__=5_1_2 ,A__=0 ,A__=1 ,A__=2 ,A__=5_0_4 ,**A__ ,): super().__init__(**A__ ,pad_token_id=A__ ,bos_token_id=A__ ,eos_token_id=A__) lowercase = hidden_size lowercase = feat_extract_norm lowercase = feat_extract_activation lowercase = list(A__) lowercase = list(A__) lowercase = list(A__) lowercase = conv_bias lowercase = num_conv_pos_embeddings lowercase = num_conv_pos_embedding_groups lowercase = len(self.conv_dim) lowercase = num_hidden_layers lowercase = intermediate_size lowercase = hidden_act lowercase = num_attention_heads lowercase = hidden_dropout lowercase = attention_dropout lowercase = activation_dropout lowercase = feat_proj_dropout lowercase = final_dropout lowercase = layerdrop lowercase = layer_norm_eps lowercase = initializer_range lowercase = vocab_size lowercase = num_clusters lowercase = do_stable_layer_norm lowercase = use_weighted_layer_sum if ( (len(self.conv_stride) != self.num_feat_extract_layers) or (len(self.conv_kernel) != self.num_feat_extract_layers) or (len(self.conv_dim) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==''' ''' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =''' f' {len(self.conv_dim)}`, `len(config.conv_stride) = {len(self.conv_stride)}`,' f' `len(config.conv_kernel) = {len(self.conv_kernel)}`.') # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 lowercase = apply_spec_augment lowercase = mask_time_prob lowercase = mask_time_length lowercase = mask_time_min_masks lowercase = mask_feature_prob lowercase = mask_feature_length lowercase = mask_feature_min_masks # parameters for pretraining with codevector quantized representations lowercase = num_codevectors_per_group lowercase = num_codevector_groups lowercase = contrastive_logits_temperature lowercase = feat_quantizer_dropout lowercase = num_negatives lowercase = codevector_dim lowercase = proj_codevector_dim lowercase = diversity_loss_weight # ctc loss lowercase = ctc_loss_reduction lowercase = ctc_zero_infinity # SequenceClassification-specific parameter. Feel free to ignore for other classes. lowercase = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. lowercase = list(A__) lowercase = list(A__) lowercase = list(A__) lowercase = xvector_output_dim @property def A__ ( self): return functools.reduce(operator.mul ,self.conv_stride ,1)
633
# Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import subprocess from packaging.version import Version, parse from accelerate.commands.config.config_args import default_config_file, load_config_from_file lowercase__ :Any = "Run commands across TPU VMs for initial setup before running `accelerate launch`." def UpperCamelCase ( lowerCAmelCase__=None ): '''simple docstring''' if subparsers is not None: lowercase = subparsers.add_parser('''tpu-config''' , description=_description ) else: lowercase = argparse.ArgumentParser('''Accelerate tpu-config command''' , description=_description ) # Core arguments lowercase = parser.add_argument_group( '''Config Arguments''' , '''Arguments that can be configured through `accelerate config`.''' ) config_args.add_argument( '''--config_file''' , type=lowerCAmelCase__ , default=lowerCAmelCase__ , help='''Path to the config file to use for accelerate.''' , ) config_args.add_argument( '''--tpu_name''' , default=lowerCAmelCase__ , help='''The name of the TPU to use. If not specified, will use the TPU specified in the config file.''' , ) config_args.add_argument( '''--tpu_zone''' , default=lowerCAmelCase__ , help='''The zone of the TPU to use. If not specified, will use the zone specified in the config file.''' , ) lowercase = parser.add_argument_group('''TPU Arguments''' , '''Arguments for options ran inside the TPU.''' ) pod_args.add_argument( '''--use_alpha''' , action='''store_true''' , help='''Whether to use `gcloud alpha` when running the TPU training script instead of `gcloud`.''' , ) pod_args.add_argument( '''--command_file''' , default=lowerCAmelCase__ , help='''The path to the file containing the commands to run on the pod on startup.''' , ) pod_args.add_argument( '''--command''' , action='''append''' , nargs='''+''' , help='''A command to run on the pod. Can be passed multiple times.''' , ) pod_args.add_argument( '''--install_accelerate''' , action='''store_true''' , help='''Whether to install accelerate on the pod. Defaults to False.''' , ) pod_args.add_argument( '''--accelerate_version''' , default='''latest''' , help='''The version of accelerate to install on the pod. If not specified, will use the latest pypi version. Specify \'dev\' to install from GitHub.''' , ) pod_args.add_argument( '''--debug''' , action='''store_true''' , help='''If set, will print the command that would be run instead of running it.''' ) if subparsers is not None: parser.set_defaults(func=lowerCAmelCase__ ) return parser def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = None # Get the default from the config file if it exists. if args.config_file is not None or os.path.isfile(lowerCAmelCase__ ): lowercase = load_config_from_file(args.config_file ) if not args.command_file and defaults.command_file is not None and not args.command: lowercase = defaults.command_file if not args.command and defaults.commands is not None: lowercase = defaults.commands if not args.tpu_name: lowercase = defaults.tpu_name if not args.tpu_zone: lowercase = defaults.tpu_zone if args.accelerate_version == "dev": lowercase = '''git+https://github.com/huggingface/accelerate.git''' elif args.accelerate_version == "latest": lowercase = '''accelerate -U''' elif isinstance(parse(args.accelerate_version ) , lowerCAmelCase__ ): lowercase = f'accelerate=={args.accelerate_version}' if not args.command_file and not args.command: raise ValueError('''You must specify either a command file or a command to run on the pod.''' ) if args.command_file: with open(args.command_file , '''r''' ) as f: lowercase = [f.read().splitlines()] # To turn list of lists into list of strings if isinstance(args.command[0] , lowerCAmelCase__ ): lowercase = [line for cmd in args.command for line in cmd] # Default to the shared folder and install accelerate lowercase = ['''cd /usr/share'''] if args.install_accelerate: new_cmd += [f'pip install {args.accelerate_version}'] new_cmd += args.command lowercase = '''; '''.join(lowerCAmelCase__ ) # Then send it to gcloud # Eventually try to use google-api-core to do this instead of subprocess lowercase = ['''gcloud'''] if args.use_alpha: cmd += ["alpha"] cmd += [ "compute", "tpus", "tpu-vm", "ssh", args.tpu_name, "--zone", args.tpu_zone, "--command", args.command, "--worker", "all", ] if args.debug: print(f'Running {" ".join(lowerCAmelCase__ )}' ) return subprocess.run(lowerCAmelCase__ ) print('''Successfully setup pod.''' ) def UpperCamelCase ( ): '''simple docstring''' lowercase = tpu_command_parser() lowercase = parser.parse_args() tpu_command_launcher(lowerCAmelCase__ )
633
1
import logging import os import threading import time try: import warnings except ImportError: lowercase__ :str = None try: import msvcrt except ImportError: lowercase__ :Tuple = None try: import fcntl except ImportError: lowercase__ :Optional[Any] = None # Backward compatibility # ------------------------------------------------ try: TimeoutError except NameError: lowercase__ :List[Any] = OSError # Data # ------------------------------------------------ lowercase__ :List[str] = [ "Timeout", "BaseFileLock", "WindowsFileLock", "UnixFileLock", "SoftFileLock", "FileLock", ] lowercase__ :Tuple = "3.0.12" lowercase__ :Tuple = None def UpperCamelCase ( ): '''simple docstring''' global _logger lowercase = _logger or logging.getLogger(__name__ ) return _logger class lowercase ( SCREAMING_SNAKE_CASE__ ): def __init__( self ,A__): lowercase = lock_file return None def __str__( self): lowercase = f'The file lock \'{self.lock_file}\' could not be acquired.' return temp class lowercase : def __init__( self ,A__): lowercase = lock return None def __enter__( self): return self.lock def __exit__( self ,A__ ,A__ ,A__): self.lock.release() return None class lowercase : def __init__( self ,A__ ,A__=-1 ,A__=None): lowercase = max_filename_length if max_filename_length is not None else 2_5_5 # Hash the filename if it's too long lowercase = self.hash_filename_if_too_long(A__ ,A__) # The path to the lock file. lowercase = lock_file # The file descriptor for the *_lock_file* as it is returned by the # os.open() function. # This file lock is only NOT None, if the object currently holds the # lock. lowercase = None # The default timeout value. lowercase = timeout # We use this lock primarily for the lock counter. lowercase = threading.Lock() # The lock counter is used for implementing the nested locking # mechanism. Whenever the lock is acquired, the counter is increased and # the lock is only released, when this value is 0 again. lowercase = 0 return None @property def A__ ( self): return self._lock_file @property def A__ ( self): return self._timeout @timeout.setter def A__ ( self ,A__): lowercase = float(A__) return None def A__ ( self): raise NotImplementedError() def A__ ( self): raise NotImplementedError() @property def A__ ( self): return self._lock_file_fd is not None def A__ ( self ,A__=None ,A__=0.05): # Use the default timeout, if no timeout is provided. if timeout is None: lowercase = self.timeout # Increment the number right at the beginning. # We can still undo it, if something fails. with self._thread_lock: self._lock_counter += 1 lowercase = id(self) lowercase = self._lock_file lowercase = time.time() try: while True: with self._thread_lock: if not self.is_locked: logger().debug(f'Attempting to acquire lock {lock_id} on {lock_filename}') self._acquire() if self.is_locked: logger().debug(f'Lock {lock_id} acquired on {lock_filename}') break elif timeout >= 0 and time.time() - start_time > timeout: logger().debug(f'Timeout on acquiring lock {lock_id} on {lock_filename}') raise Timeout(self._lock_file) else: logger().debug( f'Lock {lock_id} not acquired on {lock_filename}, waiting {poll_intervall} seconds ...') time.sleep(A__) except: # noqa # Something did go wrong, so decrement the counter. with self._thread_lock: lowercase = max(0 ,self._lock_counter - 1) raise return _Acquire_ReturnProxy(lock=self) def A__ ( self ,A__=False): with self._thread_lock: if self.is_locked: self._lock_counter -= 1 if self._lock_counter == 0 or force: lowercase = id(self) lowercase = self._lock_file logger().debug(f'Attempting to release lock {lock_id} on {lock_filename}') self._release() lowercase = 0 logger().debug(f'Lock {lock_id} released on {lock_filename}') return None def __enter__( self): self.acquire() return self def __exit__( self ,A__ ,A__ ,A__): self.release() return None def __del__( self): self.release(force=A__) return None def A__ ( self ,A__ ,A__): lowercase = os.path.basename(A__) if len(A__) > max_length and max_length > 0: lowercase = os.path.dirname(A__) lowercase = str(hash(A__)) lowercase = filename[: max_length - len(A__) - 8] + '''...''' + hashed_filename + '''.lock''' return os.path.join(A__ ,A__) else: return path class lowercase ( SCREAMING_SNAKE_CASE__ ): def __init__( self ,A__ ,A__=-1 ,A__=None): from .file_utils import relative_to_absolute_path super().__init__(A__ ,timeout=A__ ,max_filename_length=A__) lowercase = '''\\\\?\\''' + relative_to_absolute_path(self.lock_file) def A__ ( self): lowercase = os.O_RDWR | os.O_CREAT | os.O_TRUNC try: lowercase = os.open(self._lock_file ,A__) except OSError: pass else: try: msvcrt.locking(A__ ,msvcrt.LK_NBLCK ,1) except OSError: os.close(A__) else: lowercase = fd return None def A__ ( self): lowercase = self._lock_file_fd lowercase = None msvcrt.locking(A__ ,msvcrt.LK_UNLCK ,1) os.close(A__) try: os.remove(self._lock_file) # Probably another instance of the application # that acquired the file lock. except OSError: pass return None class lowercase ( SCREAMING_SNAKE_CASE__ ): def __init__( self ,A__ ,A__=-1 ,A__=None): lowercase = os.statvfs(os.path.dirname(A__)).f_namemax super().__init__(A__ ,timeout=A__ ,max_filename_length=A__) def A__ ( self): lowercase = os.O_RDWR | os.O_CREAT | os.O_TRUNC lowercase = os.open(self._lock_file ,A__) try: fcntl.flock(A__ ,fcntl.LOCK_EX | fcntl.LOCK_NB) except OSError: os.close(A__) else: lowercase = fd return None def A__ ( self): # Do not remove the lockfile: # # https://github.com/benediktschmitt/py-filelock/issues/31 # https://stackoverflow.com/questions/17708885/flock-removing-locked-file-without-race-condition lowercase = self._lock_file_fd lowercase = None fcntl.flock(A__ ,fcntl.LOCK_UN) os.close(A__) return None class lowercase ( SCREAMING_SNAKE_CASE__ ): def A__ ( self): lowercase = os.O_WRONLY | os.O_CREAT | os.O_EXCL | os.O_TRUNC try: lowercase = os.open(self._lock_file ,A__) except OSError: pass else: lowercase = fd return None def A__ ( self): os.close(self._lock_file_fd) lowercase = None try: os.remove(self._lock_file) # The file is already deleted and that's what we want. except OSError: pass return None lowercase__ :Optional[Any] = None if msvcrt: lowercase__ :Tuple = WindowsFileLock elif fcntl: lowercase__ :int = UnixFileLock else: lowercase__ :int = SoftFileLock if warnings is not None: warnings.warn("only soft file lock is available")
633
import json import multiprocessing as mp import re from collections import defaultdict from functools import partial from typing import Dict, List, Optional, Set, Tuple, Type from datasets import Dataset from datasketch import MinHash, MinHashLSH from dpu_utils.utils.iterators import ThreadedIterator from tqdm import tqdm lowercase__ :Optional[Any] = re.compile("[^A-Za-z_0-9]") # parameters used in DuplicationIndex lowercase__ :List[str] = 10 lowercase__ :Tuple = 256 def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if len(lowerCAmelCase__ ) < MIN_NUM_TOKENS: return None lowercase = MinHash(num_perm=lowerCAmelCase__ ) for token in set(lowerCAmelCase__ ): min_hash.update(token.encode() ) return min_hash def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return {t for t in NON_ALPHA.split(lowerCAmelCase__ ) if len(t.strip() ) > 0} class lowercase : def __init__( self ,*, A__ = 0.85 ,): lowercase = duplication_jaccard_threshold lowercase = NUM_PERM lowercase = MinHashLSH(threshold=self._duplication_jaccard_threshold ,num_perm=self._num_perm) lowercase = defaultdict(A__) def A__ ( self ,A__ ,A__): lowercase = self._index.query(A__) if code_key in self._index.keys: print(f'Duplicate key {code_key}') return self._index.insert(A__ ,A__) if len(A__) > 0: for base_duplicate in close_duplicates: if base_duplicate in self._duplicate_clusters: self._duplicate_clusters[base_duplicate].add(A__) break else: self._duplicate_clusters[close_duplicates[0]].add(A__) def A__ ( self): lowercase = [] for base, duplicates in self._duplicate_clusters.items(): lowercase = [base] + list(A__) # reformat the cluster to be a list of dict lowercase = [{'''base_index''': el[0], '''repo_name''': el[1], '''path''': el[2]} for el in cluster] duplicate_clusters.append(A__) return duplicate_clusters def A__ ( self ,A__): lowercase = self.get_duplicate_clusters() with open(A__ ,'''w''') as f: json.dump(A__ ,A__) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase = element lowercase = get_min_hash([t for t in NON_ALPHA.split(data['''content'''] ) if len(t.strip() ) > 0] ) if min_hash is not None: return (index, data["repo_name"], data["path"]), min_hash def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' with mp.Pool() as pool: for data in pool.imap_unordered( _compute_min_hash , ThreadedIterator(lowerCAmelCase__ , max_queue_size=1_0000 ) , chunksize=100 , ): if data is not None: yield data def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = DuplicationIndex(duplication_jaccard_threshold=lowerCAmelCase__ ) for filename, min_hash in tqdm(ThreadedIterator(minhash_iter(enumerate(lowerCAmelCase__ ) ) , max_queue_size=100 ) ): di.add(lowerCAmelCase__ , lowerCAmelCase__ ) # Returns a List[Cluster] where Cluster is List[str] with the filenames. return di.get_duplicate_clusters() def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = get_tokens(lowerCAmelCase__ ) lowercase = get_tokens(lowerCAmelCase__ ) return len(tokensa & tokensa ) / len(tokensa | tokensa ) lowercase__ :List[Any] = None def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = [] for elementa in cluster: lowercase = _shared_dataset[elementa['''base_index''']]['''content'''] for elementa in extremes: lowercase = _shared_dataset[elementa['''base_index''']]['''content'''] if jaccard_similarity(lowerCAmelCase__ , lowerCAmelCase__ ) >= jaccard_threshold: elementa["copies"] += 1 break else: lowercase = 1 extremes.append(lowerCAmelCase__ ) return extremes def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' global _shared_dataset lowercase = dataset lowercase = [] lowercase = partial(_find_cluster_extremes_shared , jaccard_threshold=lowerCAmelCase__ ) with mp.Pool() as pool: for extremes in tqdm( pool.imap_unordered( lowerCAmelCase__ , lowerCAmelCase__ , ) , total=len(lowerCAmelCase__ ) , ): extremes_list.append(lowerCAmelCase__ ) return extremes_list def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ = 0.85 ): '''simple docstring''' lowercase = make_duplicate_clusters(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase = {x['''base_index'''] for cluster in duplicate_clusters for x in cluster} lowercase = {} lowercase = find_extremes(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) for extremes in extremes_clusters: for element in extremes: lowercase = element lowercase = duplicate_indices - set(extreme_dict.keys() ) lowercase = dataset.filter(lambda lowerCAmelCase__ , lowerCAmelCase__ : idx not in remove_indices , with_indices=lowerCAmelCase__ ) # update duplicate_clusters for cluster in duplicate_clusters: for element in cluster: lowercase = element['''base_index'''] in extreme_dict if element["is_extreme"]: lowercase = extreme_dict[element['''base_index''']]['''copies'''] print(f'Original dataset size: {len(lowerCAmelCase__ )}' ) print(f'Number of duplicate clusters: {len(lowerCAmelCase__ )}' ) print(f'Files in duplicate cluster: {len(lowerCAmelCase__ )}' ) print(f'Unique files in duplicate cluster: {len(lowerCAmelCase__ )}' ) print(f'Filtered dataset size: {len(lowerCAmelCase__ )}' ) return ds_filter, duplicate_clusters
633
1
import numpy as np import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModelWithProjection, PreTrainedModel from ...utils import logging lowercase__ :Union[str, Any] = logging.get_logger(__name__) class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Union[str, Any] =CLIPConfig lowercase_ : str =['''CLIPEncoderLayer'''] def __init__( self ,A__): super().__init__(A__) lowercase = CLIPVisionModelWithProjection(config.vision_config) lowercase = nn.Linear(config.vision_config.projection_dim ,1) lowercase = nn.Linear(config.vision_config.projection_dim ,1) @torch.no_grad() def A__ ( self ,A__ ,A__ ,A__=0.5 ,A__=0.5): lowercase = self.vision_model(A__)[0] lowercase = self.p_head(A__) lowercase = nsfw_detected.flatten() lowercase = nsfw_detected > p_threshold lowercase = nsfw_detected.tolist() if any(A__): logger.warning( '''Potential NSFW content was detected in one or more images. A black image will be returned instead.''' ''' Try again with a different prompt and/or seed.''') for idx, nsfw_detected_ in enumerate(A__): if nsfw_detected_: lowercase = np.zeros(images[idx].shape) lowercase = self.w_head(A__) lowercase = watermark_detected.flatten() lowercase = watermark_detected > w_threshold lowercase = watermark_detected.tolist() if any(A__): logger.warning( '''Potential watermarked content was detected in one or more images. A black image will be returned instead.''' ''' Try again with a different prompt and/or seed.''') for idx, watermark_detected_ in enumerate(A__): if watermark_detected_: lowercase = np.zeros(images[idx].shape) return images, nsfw_detected, watermark_detected
633
import numpy as np import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModelWithProjection, PreTrainedModel from ...utils import logging lowercase__ :Union[str, Any] = logging.get_logger(__name__) class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Union[str, Any] =CLIPConfig lowercase_ : str =['''CLIPEncoderLayer'''] def __init__( self ,A__): super().__init__(A__) lowercase = CLIPVisionModelWithProjection(config.vision_config) lowercase = nn.Linear(config.vision_config.projection_dim ,1) lowercase = nn.Linear(config.vision_config.projection_dim ,1) @torch.no_grad() def A__ ( self ,A__ ,A__ ,A__=0.5 ,A__=0.5): lowercase = self.vision_model(A__)[0] lowercase = self.p_head(A__) lowercase = nsfw_detected.flatten() lowercase = nsfw_detected > p_threshold lowercase = nsfw_detected.tolist() if any(A__): logger.warning( '''Potential NSFW content was detected in one or more images. A black image will be returned instead.''' ''' Try again with a different prompt and/or seed.''') for idx, nsfw_detected_ in enumerate(A__): if nsfw_detected_: lowercase = np.zeros(images[idx].shape) lowercase = self.w_head(A__) lowercase = watermark_detected.flatten() lowercase = watermark_detected > w_threshold lowercase = watermark_detected.tolist() if any(A__): logger.warning( '''Potential watermarked content was detected in one or more images. A black image will be returned instead.''' ''' Try again with a different prompt and/or seed.''') for idx, watermark_detected_ in enumerate(A__): if watermark_detected_: lowercase = np.zeros(images[idx].shape) return images, nsfw_detected, watermark_detected
633
1
from __future__ import annotations class lowercase : def __init__( self ,A__): lowercase = order # a_{0} ... a_{k} lowercase = [1.0] + [0.0] * order # b_{0} ... b_{k} lowercase = [1.0] + [0.0] * order # x[n-1] ... x[n-k] lowercase = [0.0] * self.order # y[n-1] ... y[n-k] lowercase = [0.0] * self.order def A__ ( self ,A__ ,A__): if len(A__) < self.order: lowercase = [1.0, *a_coeffs] if len(A__) != self.order + 1: lowercase = ( f'Expected a_coeffs to have {self.order + 1} elements ' f'for {self.order}-order filter, got {len(A__)}' ) raise ValueError(A__) if len(A__) != self.order + 1: lowercase = ( f'Expected b_coeffs to have {self.order + 1} elements ' f'for {self.order}-order filter, got {len(A__)}' ) raise ValueError(A__) lowercase = a_coeffs lowercase = b_coeffs def A__ ( self ,A__): lowercase = 0.0 # Start at index 1 and do index 0 at the end. for i in range(1 ,self.order + 1): result += ( self.b_coeffs[i] * self.input_history[i - 1] - self.a_coeffs[i] * self.output_history[i - 1] ) lowercase = (result + self.b_coeffs[0] * sample) / self.a_coeffs[0] lowercase = self.input_history[:-1] lowercase = self.output_history[:-1] lowercase = sample lowercase = result return result
633
class lowercase : def __init__( self ,A__): lowercase = val lowercase = None lowercase = None def A__ ( self ,A__): if self.val: if val < self.val: if self.left is None: lowercase = Node(A__) else: self.left.insert(A__) elif val > self.val: if self.right is None: lowercase = Node(A__) else: self.right.insert(A__) else: lowercase = val def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' # Recursive traversal if root: inorder(root.left , lowerCAmelCase__ ) res.append(root.val ) inorder(root.right , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' # Build BST if len(lowerCAmelCase__ ) == 0: return arr lowercase = Node(arr[0] ) for i in range(1 , len(lowerCAmelCase__ ) ): root.insert(arr[i] ) # Traverse BST in order. lowercase = [] inorder(lowerCAmelCase__ , lowerCAmelCase__ ) return res if __name__ == "__main__": print(tree_sort([10, 1, 3, 2, 9, 14, 13]))
633
1
# Lint as: python3 import itertools import os import re lowercase__ :List[str] = re.compile(r"([A-Z]+)([A-Z][a-z])") lowercase__ :int = re.compile(r"([a-z\d])([A-Z])") lowercase__ :List[str] = re.compile(r"(?<!_)_(?!_)") lowercase__ :List[Any] = re.compile(r"(_{2,})") lowercase__ :List[str] = r"^\w+(\.\w+)*$" lowercase__ :Optional[Any] = r"<>:/\|?*" def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = _uppercase_uppercase_re.sub(R'''\1_\2''' , lowerCAmelCase__ ) lowercase = _lowercase_uppercase_re.sub(R'''\1_\2''' , lowerCAmelCase__ ) return name.lower() def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = _single_underscore_re.split(lowerCAmelCase__ ) lowercase = [_multiple_underscores_re.split(lowerCAmelCase__ ) for n in name] return "".join(n.capitalize() for n in itertools.chain.from_iterable(lowerCAmelCase__ ) if n != '''''' ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if os.path.basename(lowerCAmelCase__ ) != name: raise ValueError(f'Should be a dataset name, not a path: {name}' ) return camelcase_to_snakecase(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' if os.path.basename(lowerCAmelCase__ ) != name: raise ValueError(f'Should be a dataset name, not a path: {name}' ) if not re.match(_split_re , lowerCAmelCase__ ): raise ValueError(f'Split name should match \'{_split_re}\'\' but got \'{split}\'.' ) return f'{filename_prefix_for_name(lowerCAmelCase__ )}-{split}' def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None ): '''simple docstring''' lowercase = filename_prefix_for_split(lowerCAmelCase__ , lowerCAmelCase__ ) if filetype_suffix: prefix += f'.{filetype_suffix}' lowercase = os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) return f'{filepath}*' def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None ): '''simple docstring''' lowercase = filename_prefix_for_split(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase = os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) if shard_lengths: lowercase = len(lowerCAmelCase__ ) lowercase = [f'{prefix}-{shard_id:05d}-of-{num_shards:05d}' for shard_id in range(lowerCAmelCase__ )] if filetype_suffix: lowercase = [filename + f'.{filetype_suffix}' for filename in filenames] return filenames else: lowercase = prefix if filetype_suffix: filename += f'.{filetype_suffix}' return [filename]
633
import os def UpperCamelCase ( lowerCAmelCase__ = "input.txt" ): '''simple docstring''' with open(os.path.join(os.path.dirname(lowerCAmelCase__ ) , lowerCAmelCase__ ) ) as input_file: lowercase = [ [int(lowerCAmelCase__ ) for element in line.split(''',''' )] for line in input_file.readlines() ] lowercase = len(lowerCAmelCase__ ) lowercase = len(matrix[0] ) lowercase = [[-1 for _ in range(lowerCAmelCase__ )] for _ in range(lowerCAmelCase__ )] for i in range(lowerCAmelCase__ ): lowercase = matrix[i][0] for j in range(1 , lowerCAmelCase__ ): for i in range(lowerCAmelCase__ ): lowercase = minimal_path_sums[i][j - 1] + matrix[i][j] for i in range(1 , lowerCAmelCase__ ): lowercase = min( minimal_path_sums[i][j] , minimal_path_sums[i - 1][j] + matrix[i][j] ) for i in range(rows - 2 , -1 , -1 ): lowercase = min( minimal_path_sums[i][j] , minimal_path_sums[i + 1][j] + matrix[i][j] ) return min(minimal_path_sums_row[-1] for minimal_path_sums_row in minimal_path_sums ) if __name__ == "__main__": print(F'{solution() = }')
633
1
import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : List[str] =(UniPCMultistepScheduler,) lowercase_ : Tuple =(('''num_inference_steps''', 25),) def A__ ( self ,**A__): lowercase = { '''num_train_timesteps''': 1_0_0_0, '''beta_start''': 0.0001, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', '''solver_order''': 2, '''solver_type''': '''bh2''', } config.update(**A__) return config def A__ ( self ,A__=0 ,**A__): lowercase = dict(self.forward_default_kwargs) lowercase = kwargs.pop('''num_inference_steps''' ,A__) lowercase = self.dummy_sample lowercase = 0.1 * sample lowercase = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) scheduler.set_timesteps(A__) # copy over dummy past residuals lowercase = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A__) lowercase = scheduler_class.from_pretrained(A__) new_scheduler.set_timesteps(A__) # copy over dummy past residuals lowercase = dummy_past_residuals[: new_scheduler.config.solver_order] lowercase , lowercase = sample, sample for t in range(A__ ,time_step + scheduler.config.solver_order + 1): lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = new_scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1E-5, "Scheduler outputs are not identical" def A__ ( self ,A__=0 ,**A__): lowercase = dict(self.forward_default_kwargs) lowercase = kwargs.pop('''num_inference_steps''' ,A__) lowercase = self.dummy_sample lowercase = 0.1 * sample lowercase = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config() lowercase = scheduler_class(**A__) scheduler.set_timesteps(A__) # copy over dummy past residuals (must be after setting timesteps) lowercase = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A__) lowercase = scheduler_class.from_pretrained(A__) # copy over dummy past residuals new_scheduler.set_timesteps(A__) # copy over dummy past residual (must be after setting timesteps) lowercase = dummy_past_residuals[: new_scheduler.config.solver_order] lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = new_scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1E-5, "Scheduler outputs are not identical" def A__ ( self ,A__=None ,**A__): if scheduler is None: lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) lowercase = 1_0 lowercase = self.dummy_model() lowercase = self.dummy_sample_deter scheduler.set_timesteps(A__) for i, t in enumerate(scheduler.timesteps): lowercase = model(A__ ,A__) lowercase = scheduler.step(A__ ,A__ ,A__).prev_sample return sample def A__ ( self): lowercase = dict(self.forward_default_kwargs) lowercase = kwargs.pop('''num_inference_steps''' ,A__) for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config() lowercase = scheduler_class(**A__) lowercase = self.dummy_sample lowercase = 0.1 * sample if num_inference_steps is not None and hasattr(A__ ,'''set_timesteps'''): scheduler.set_timesteps(A__) elif num_inference_steps is not None and not hasattr(A__ ,'''set_timesteps'''): lowercase = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) lowercase = [residual + 0.2, residual + 0.15, residual + 0.10] lowercase = dummy_past_residuals[: scheduler.config.solver_order] lowercase = scheduler.timesteps[5] lowercase = scheduler.timesteps[6] lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample self.assertEqual(output_a.shape ,sample.shape) self.assertEqual(output_a.shape ,output_a.shape) def A__ ( self): # make sure that iterating over schedulers with same config names gives same results # for defaults lowercase = UniPCMultistepScheduler(**self.get_scheduler_config()) lowercase = self.full_loop(scheduler=A__) lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.2464) < 1E-3 lowercase = DPMSolverSinglestepScheduler.from_config(scheduler.config) lowercase = DEISMultistepScheduler.from_config(scheduler.config) lowercase = DPMSolverMultistepScheduler.from_config(scheduler.config) lowercase = UniPCMultistepScheduler.from_config(scheduler.config) lowercase = self.full_loop(scheduler=A__) lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.2464) < 1E-3 def A__ ( self): for timesteps in [2_5, 5_0, 1_0_0, 9_9_9, 1_0_0_0]: self.check_over_configs(num_train_timesteps=A__) def A__ ( self): self.check_over_configs(thresholding=A__) for order in [1, 2, 3]: for solver_type in ["bh1", "bh2"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=A__ ,prediction_type=A__ ,sample_max_value=A__ ,solver_order=A__ ,solver_type=A__ ,) def A__ ( self): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=A__) def A__ ( self): for solver_type in ["bh1", "bh2"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=A__ ,solver_type=A__ ,prediction_type=A__ ,) lowercase = self.full_loop( solver_order=A__ ,solver_type=A__ ,prediction_type=A__ ,) assert not torch.isnan(A__).any(), "Samples have nan numbers" def A__ ( self): self.check_over_configs(lower_order_final=A__) self.check_over_configs(lower_order_final=A__) def A__ ( self): for num_inference_steps in [1, 2, 3, 5, 1_0, 5_0, 1_0_0, 9_9_9, 1_0_0_0]: self.check_over_forward(num_inference_steps=A__ ,time_step=0) def A__ ( self): lowercase = self.full_loop() lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.2464) < 1E-3 def A__ ( self): lowercase = self.full_loop(prediction_type='''v_prediction''') lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.1014) < 1E-3 def A__ ( self): lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config(thresholding=A__ ,dynamic_thresholding_ratio=0) lowercase = scheduler_class(**A__) lowercase = 1_0 lowercase = self.dummy_model() lowercase = self.dummy_sample_deter.half() scheduler.set_timesteps(A__) for i, t in enumerate(scheduler.timesteps): lowercase = model(A__ ,A__) lowercase = scheduler.step(A__ ,A__ ,A__).prev_sample assert sample.dtype == torch.floataa def A__ ( self ,**A__): for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) scheduler.set_timesteps(scheduler.config.num_train_timesteps) assert len(scheduler.timesteps.unique()) == scheduler.num_inference_steps
633
from __future__ import annotations def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if len(lowerCAmelCase__ ) < 2: raise ValueError('''Monogons and Digons are not polygons in the Euclidean space''' ) if any(i <= 0 for i in nums ): raise ValueError('''All values must be greater than 0''' ) lowercase = nums.copy() copy_nums.sort() return copy_nums[-1] < sum(copy_nums[:-1] ) if __name__ == "__main__": import doctest doctest.testmod()
633
1
import gc import importlib.metadata import tempfile import unittest from packaging import version from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoTokenizer, BitsAndBytesConfig, pipeline, ) from transformers.testing_utils import ( is_torch_available, require_accelerate, require_bitsandbytes, require_torch, require_torch_gpu, require_torch_multi_gpu, slow, ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if model.config.model_type == "gpt2": return model.transformer.h[0].mlp.c_fc return model.transformer.h[0].mlp.dense_ah_to_h if is_torch_available(): import torch import torch.nn as nn class lowercase ( nn.Module ): def __init__( self ,A__ ,A__): super().__init__() lowercase = module lowercase = nn.Sequential( nn.Linear(module.in_features ,A__ ,bias=A__) ,nn.Linear(A__ ,module.out_features ,bias=A__) ,) lowercase = (2.0 / (5 * min(module.in_features ,module.out_features))) ** 0.5 nn.init.normal_(self.adapter[0].weight ,std=A__) nn.init.zeros_(self.adapter[1].weight) self.adapter.to(module.weight.device) def A__ ( self ,A__ ,*A__ ,**A__): return self.module(A__ ,*A__ ,**A__) + self.adapter(A__) @require_bitsandbytes @require_accelerate @require_torch @require_torch_gpu @slow class lowercase ( unittest.TestCase ): # We keep the constants inside the init function and model loading inside setUp function # We need to test on relatively large models (aka >1b parameters otherwise the quantiztion may not work as expected) # Therefore here we use only bloom-1b3 to test our module lowercase_ : Any ='''bigscience/bloom-1b7''' # Constant values lowercase_ : str =2.109_659_552_692_574 lowercase_ : str ='''Hello my name is''' lowercase_ : Union[str, Any] =set() EXPECTED_OUTPUTS.add('''Hello my name is John and I am a professional photographer. I''' ) EXPECTED_OUTPUTS.add('''Hello my name is John.\nI am a friend of your father.\n''' ) EXPECTED_OUTPUTS.add('''Hello my name is John Doe, I am a student at the University''' ) lowercase_ : Optional[Any] =10 def A__ ( self): # Models and tokenizer lowercase = AutoTokenizer.from_pretrained(self.model_name) class lowercase ( SCREAMING_SNAKE_CASE__ ): def A__ ( self): super().setUp() # Models and tokenizer lowercase = AutoModelForCausalLM.from_pretrained( self.model_name ,torch_dtype=torch.floataa ,device_map='''auto''') lowercase = AutoModelForCausalLM.from_pretrained(self.model_name ,load_in_abit=A__ ,device_map='''auto''') def A__ ( self): del self.model_fpaa del self.model_abit gc.collect() torch.cuda.empty_cache() def A__ ( self): lowercase = self.model_abit.config self.assertTrue(hasattr(A__ ,'''quantization_config''')) lowercase = config.to_dict() lowercase = config.to_diff_dict() lowercase = config.to_json_string() def A__ ( self): from bitsandbytes.nn import Paramsabit lowercase = self.model_fpaa.get_memory_footprint() lowercase = self.model_abit.get_memory_footprint() self.assertAlmostEqual(mem_fpaa / mem_abit ,self.EXPECTED_RELATIVE_DIFFERENCE) lowercase = get_some_linear_layer(self.model_abit) self.assertTrue(linear.weight.__class__ == Paramsabit) def A__ ( self): from transformers import TaPreTrainedModel self.model_fpaa.get_memory_footprint() self.model_abit.get_memory_footprint() for name, module in self.model_abit.named_modules(): if isinstance(A__ ,torch.nn.Linear): if name not in ["lm_head"] + TaPreTrainedModel._keep_in_fpaa_modules: # 4-bit parameters are packed in uint8 variables self.assertTrue(module.weight.dtype == torch.uinta) def A__ ( self): lowercase = self.tokenizer(self.input_text ,return_tensors='''pt''') lowercase = self.model_abit.generate(input_ids=encoded_input['''input_ids'''].to(0) ,max_new_tokens=1_0) self.assertIn(self.tokenizer.decode(output_sequences[0] ,skip_special_tokens=A__) ,self.EXPECTED_OUTPUTS) def A__ ( self): lowercase = BitsAndBytesConfig() lowercase = True lowercase = AutoModelForCausalLM.from_pretrained( self.model_name ,quantization_config=A__ ,device_map='''auto''') lowercase = self.tokenizer(self.input_text ,return_tensors='''pt''') lowercase = model_abit_from_config.generate( input_ids=encoded_input['''input_ids'''].to(0) ,max_new_tokens=1_0) self.assertIn(self.tokenizer.decode(output_sequences[0] ,skip_special_tokens=A__) ,self.EXPECTED_OUTPUTS) def A__ ( self): with self.assertRaises(A__), tempfile.TemporaryDirectory() as tmpdirname: self.model_abit.save_pretrained(A__) def A__ ( self): lowercase = BitsAndBytesConfig() with self.assertRaises(A__): lowercase = AutoModelForCausalLM.from_pretrained( self.model_name ,quantization_config=A__ ,load_in_abit=A__ ,device_map='''auto''' ,bnb_abit_quant_type='''nf4''' ,) def A__ ( self): with self.assertRaises(A__): # Tries with `str` self.model_abit.to('''cpu''') with self.assertRaises(A__): # Tries with a `dtype`` self.model_abit.to(torch.floataa) with self.assertRaises(A__): # Tries with a `device` self.model_abit.to(torch.device('''cuda:0''')) with self.assertRaises(A__): # Tries with a `device` self.model_abit.float() with self.assertRaises(A__): # Tries with a `device` self.model_abit.half() # Test if we did not break anything lowercase = self.tokenizer(self.input_text ,return_tensors='''pt''') lowercase = self.model_fpaa.to(torch.floataa) lowercase = self.model_fpaa.generate(input_ids=encoded_input['''input_ids'''].to(0) ,max_new_tokens=1_0) # Check this does not throw an error lowercase = self.model_fpaa.to('''cpu''') # Check this does not throw an error lowercase = self.model_fpaa.half() # Check this does not throw an error lowercase = self.model_fpaa.float() def A__ ( self): lowercase = AutoModelForSeqaSeqLM.from_pretrained('''t5-small''' ,load_in_abit=A__ ,device_map='''auto''') self.assertTrue(model.decoder.block[0].layer[2].DenseReluDense.wo.weight.dtype == torch.floataa) @require_bitsandbytes @require_accelerate @require_torch @require_torch_gpu @slow class lowercase ( unittest.TestCase ): @classmethod def A__ ( cls): lowercase = '''t5-small''' lowercase = '''google/flan-t5-small''' # flan-t5 uses dense-act instead of dense-relu-dense lowercase = AutoTokenizer.from_pretrained(cls.model_name) lowercase = '''Translate in German: Hello, my dog is cute''' def A__ ( self): gc.collect() torch.cuda.empty_cache() def A__ ( self): from transformers import TaForConditionalGeneration lowercase = TaForConditionalGeneration._keep_in_fpaa_modules lowercase = None # test with `t5-small` lowercase = TaForConditionalGeneration.from_pretrained(self.model_name ,load_in_abit=A__ ,device_map='''auto''') lowercase = self.tokenizer(self.input_text ,return_tensors='''pt''').to(0) lowercase = model.generate(**A__) # test with `flan-t5-small` lowercase = TaForConditionalGeneration.from_pretrained( self.dense_act_model_name ,load_in_abit=A__ ,device_map='''auto''') lowercase = self.tokenizer(self.input_text ,return_tensors='''pt''').to(0) lowercase = model.generate(**A__) lowercase = modules def A__ ( self): import bitsandbytes as bnb from transformers import TaForConditionalGeneration # test with `t5-small` lowercase = TaForConditionalGeneration.from_pretrained(self.model_name ,load_in_abit=A__ ,device_map='''auto''') # there was a bug with decoders - this test checks that it is fixed self.assertTrue(isinstance(model.decoder.block[0].layer[0].SelfAttention.q ,bnb.nn.Linearabit)) lowercase = self.tokenizer(self.input_text ,return_tensors='''pt''').to(0) lowercase = model.generate(**A__) # test with `flan-t5-small` lowercase = TaForConditionalGeneration.from_pretrained( self.dense_act_model_name ,load_in_abit=A__ ,device_map='''auto''') lowercase = self.tokenizer(self.input_text ,return_tensors='''pt''').to(0) lowercase = model.generate(**A__) class lowercase ( SCREAMING_SNAKE_CASE__ ): def A__ ( self): super().setUp() # model_name lowercase = '''bigscience/bloom-560m''' lowercase = '''t5-small''' # Different types of model lowercase = AutoModel.from_pretrained(self.model_name ,load_in_abit=A__ ,device_map='''auto''') # Sequence classification model lowercase = AutoModelForSequenceClassification.from_pretrained( self.model_name ,load_in_abit=A__ ,device_map='''auto''') # CausalLM model lowercase = AutoModelForCausalLM.from_pretrained(self.model_name ,load_in_abit=A__ ,device_map='''auto''') # Seq2seq model lowercase = AutoModelForSeqaSeqLM.from_pretrained( self.seq_to_seq_name ,load_in_abit=A__ ,device_map='''auto''') def A__ ( self): del self.base_model del self.sequence_model del self.model_abit del self.seq_to_seq_model gc.collect() torch.cuda.empty_cache() def A__ ( self): from bitsandbytes.nn import Paramsabit self.assertTrue(self.base_model.h[-1].mlp.dense_ah_to_h.weight.__class__ == Paramsabit) # Other heads should be nn.Parameter self.assertTrue(self.model_abit.lm_head.weight.__class__ == torch.nn.Parameter) self.assertTrue(self.sequence_model.score.weight.__class__ == torch.nn.Parameter) self.assertTrue(self.seq_to_seq_model.lm_head.weight.__class__ == torch.nn.Parameter) class lowercase ( SCREAMING_SNAKE_CASE__ ): def A__ ( self): super().setUp() def A__ ( self): del self.pipe gc.collect() torch.cuda.empty_cache() def A__ ( self): lowercase = pipeline( '''text-generation''' ,model=self.model_name ,model_kwargs={'''device_map''': '''auto''', '''load_in_4bit''': True, '''torch_dtype''': torch.floataa} ,max_new_tokens=self.MAX_NEW_TOKENS ,) # Real second forward pass lowercase = self.pipe(self.input_text) self.assertIn(pipeline_output[0]['''generated_text'''] ,self.EXPECTED_OUTPUTS) @require_torch_multi_gpu class lowercase ( SCREAMING_SNAKE_CASE__ ): def A__ ( self): super().setUp() def A__ ( self): lowercase = AutoModelForCausalLM.from_pretrained( self.model_name ,load_in_abit=A__ ,device_map='''balanced''') # Check correct device map self.assertEqual(set(model_parallel.hf_device_map.values()) ,{0, 1}) # Check that inference pass works on the model lowercase = self.tokenizer(self.input_text ,return_tensors='''pt''') # Second real batch lowercase = model_parallel.generate(input_ids=encoded_input['''input_ids'''].to(0) ,max_new_tokens=1_0) self.assertIn(self.tokenizer.decode(output_parallel[0] ,skip_special_tokens=A__) ,self.EXPECTED_OUTPUTS) class lowercase ( SCREAMING_SNAKE_CASE__ ): def A__ ( self): lowercase = '''facebook/opt-350m''' super().setUp() def A__ ( self): if version.parse(importlib.metadata.version('''bitsandbytes''')) < version.parse('''0.37.0'''): return # Step 1: freeze all parameters lowercase = AutoModelForCausalLM.from_pretrained(self.model_name ,load_in_abit=A__) self.assertEqual(set(model.hf_device_map.values()) ,{torch.cuda.current_device()}) for param in model.parameters(): lowercase = False # freeze the model - train adapters later if param.ndim == 1: # cast the small parameters (e.g. layernorm) to fp32 for stability lowercase = param.data.to(torch.floataa) # Step 2: add adapters for _, module in model.named_modules(): if "OPTAttention" in repr(type(A__)): lowercase = LoRALayer(module.q_proj ,rank=1_6) lowercase = LoRALayer(module.k_proj ,rank=1_6) lowercase = LoRALayer(module.v_proj ,rank=1_6) # Step 3: dummy batch lowercase = self.tokenizer('''Test batch ''' ,return_tensors='''pt''').to(0) # Step 4: Check if the gradient is not None with torch.cuda.amp.autocast(): lowercase = model.forward(**A__) out.logits.norm().backward() for module in model.modules(): if isinstance(A__ ,A__): self.assertTrue(module.adapter[1].weight.grad is not None) self.assertTrue(module.adapter[1].weight.grad.norm().item() > 0) elif isinstance(A__ ,nn.Embedding): self.assertTrue(module.weight.grad is None) class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Any ='''gpt2-xl''' lowercase_ : Union[str, Any] =3.3_191_854_854_152_187
633
def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' return int((input_a, input_a).count(0 ) == 0 ) def UpperCamelCase ( ): '''simple docstring''' assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_gate(1 , 0 ) == 0 assert and_gate(1 , 1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
633
1
from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ :List[Any] = logging.get_logger(__name__) lowercase__ :Optional[int] = { "facebook/dpr-ctx_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/config.json" ), "facebook/dpr-question_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/config.json" ), "facebook/dpr-reader-single-nq-base": ( "https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/config.json" ), "facebook/dpr-ctx_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/config.json" ), "facebook/dpr-question_encoder-multiset-base": ( "https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/config.json" ), "facebook/dpr-reader-multiset-base": ( "https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/config.json" ), } class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Tuple ='''dpr''' def __init__( self ,A__=3_0_5_2_2 ,A__=7_6_8 ,A__=1_2 ,A__=1_2 ,A__=3_0_7_2 ,A__="gelu" ,A__=0.1 ,A__=0.1 ,A__=5_1_2 ,A__=2 ,A__=0.02 ,A__=1E-12 ,A__=0 ,A__="absolute" ,A__ = 0 ,**A__ ,): super().__init__(pad_token_id=A__ ,**A__) lowercase = vocab_size lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_attention_heads lowercase = hidden_act lowercase = intermediate_size lowercase = hidden_dropout_prob lowercase = attention_probs_dropout_prob lowercase = max_position_embeddings lowercase = type_vocab_size lowercase = initializer_range lowercase = layer_norm_eps lowercase = projection_dim lowercase = position_embedding_type
633
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowercase__ :Tuple = { "configuration_biogpt": ["BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BioGptConfig"], "tokenization_biogpt": ["BioGptTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :Union[str, Any] = [ "BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST", "BioGptForCausalLM", "BioGptForTokenClassification", "BioGptForSequenceClassification", "BioGptModel", "BioGptPreTrainedModel", ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys lowercase__ :Any = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
633
1
from __future__ import annotations from math import pi def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' if (inductance, frequency, reactance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if inductance < 0: raise ValueError('''Inductance cannot be negative''' ) if frequency < 0: raise ValueError('''Frequency cannot be negative''' ) if reactance < 0: raise ValueError('''Inductive reactance cannot be negative''' ) if inductance == 0: return {"inductance": reactance / (2 * pi * frequency)} elif frequency == 0: return {"frequency": reactance / (2 * pi * inductance)} elif reactance == 0: return {"reactance": 2 * pi * frequency * inductance} else: raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
633
import logging from transformers import PretrainedConfig lowercase__ :int = logging.getLogger(__name__) lowercase__ :Dict = { "bertabs-finetuned-cnndm": "https://huggingface.co/remi/bertabs-finetuned-cnndm-extractive-abstractive-summarization/resolve/main/config.json", } class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Optional[int] ='''bertabs''' def __init__( self ,A__=3_0_5_2_2 ,A__=5_1_2 ,A__=6 ,A__=5_1_2 ,A__=8 ,A__=5_1_2 ,A__=0.2 ,A__=6 ,A__=7_6_8 ,A__=8 ,A__=2_0_4_8 ,A__=0.2 ,**A__ ,): super().__init__(**A__) lowercase = vocab_size lowercase = max_pos lowercase = enc_layers lowercase = enc_hidden_size lowercase = enc_heads lowercase = enc_ff_size lowercase = enc_dropout lowercase = dec_layers lowercase = dec_hidden_size lowercase = dec_heads lowercase = dec_ff_size lowercase = dec_dropout
633
1
import datasets from .evaluate import evaluate lowercase__ :Tuple = "\\n@article{hendrycks2021cuad,\n title={CUAD: An Expert-Annotated NLP Dataset for Legal Contract Review},\n author={Dan Hendrycks and Collin Burns and Anya Chen and Spencer Ball},\n journal={arXiv preprint arXiv:2103.06268},\n year={2021}\n}\n" lowercase__ :Tuple = "\nThis metric wrap the official scoring script for version 1 of the Contract\nUnderstanding Atticus Dataset (CUAD).\nContract Understanding Atticus Dataset (CUAD) v1 is a corpus of more than 13,000 labels in 510\ncommercial legal contracts that have been manually labeled to identify 41 categories of important\nclauses that lawyers look for when reviewing contracts in connection with corporate transactions.\n" lowercase__ :Dict = "\nComputes CUAD scores (EM, F1, AUPR, Precision@80%Recall, and Precision@90%Recall).\nArgs:\n predictions: List of question-answers dictionaries with the following key-values:\n - 'id': id of the question-answer pair as given in the references (see below)\n - 'prediction_text': list of possible texts for the answer, as a list of strings\n depending on a threshold on the confidence probability of each prediction.\n references: List of question-answers dictionaries with the following key-values:\n - 'id': id of the question-answer pair (see above),\n - 'answers': a Dict in the CUAD dataset format\n {\n 'text': list of possible texts for the answer, as a list of strings\n 'answer_start': list of start positions for the answer, as a list of ints\n }\n Note that answer_start values are not taken into account to compute the metric.\nReturns:\n 'exact_match': Exact match (the normalized answer exactly match the gold answer)\n 'f1': The F-score of predicted tokens versus the gold answer\n 'aupr': Area Under the Precision-Recall curve\n 'prec_at_80_recall': Precision at 80% recall\n 'prec_at_90_recall': Precision at 90% recall\nExamples:\n >>> predictions = [{'prediction_text': ['The seller:', 'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.'], 'id': 'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties'}]\n >>> references = [{'answers': {'answer_start': [143, 49], 'text': ['The seller:', 'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.']}, 'id': 'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties'}]\n >>> cuad_metric = datasets.load_metric(\"cuad\")\n >>> results = cuad_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'exact_match': 100.0, 'f1': 100.0, 'aupr': 0.0, 'prec_at_80_recall': 1.0, 'prec_at_90_recall': 1.0}\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase ( datasets.Metric ): def A__ ( self): return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': { '''id''': datasets.Value('''string'''), '''prediction_text''': datasets.features.Sequence(datasets.Value('''string''')), }, '''references''': { '''id''': datasets.Value('''string'''), '''answers''': datasets.features.Sequence( { '''text''': datasets.Value('''string'''), '''answer_start''': datasets.Value('''int32'''), }), }, }) ,codebase_urls=['''https://www.atticusprojectai.org/cuad'''] ,reference_urls=['''https://www.atticusprojectai.org/cuad'''] ,) def A__ ( self ,A__ ,A__): lowercase = {prediction['''id''']: prediction['''prediction_text'''] for prediction in predictions} lowercase = [ { '''paragraphs''': [ { '''qas''': [ { '''answers''': [{'''text''': answer_text} for answer_text in ref['''answers''']['''text''']], '''id''': ref['''id'''], } for ref in references ] } ] } ] lowercase = evaluate(dataset=A__ ,predictions=A__) return score
633
import warnings from diffusers import StableDiffusionInpaintPipeline as StableDiffusionInpaintPipeline # noqa F401 warnings.warn( "The `inpainting.py` script is outdated. Please use directly `from diffusers import" " StableDiffusionInpaintPipeline` instead." )
633
1
import torch from diffusers import CMStochasticIterativeScheduler from .test_schedulers import SchedulerCommonTest class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : List[str] =(CMStochasticIterativeScheduler,) lowercase_ : str =10 def A__ ( self ,**A__): lowercase = { '''num_train_timesteps''': 2_0_1, '''sigma_min''': 0.002, '''sigma_max''': 80.0, } config.update(**A__) return config def A__ ( self): lowercase = 1_0 lowercase = self.get_scheduler_config() lowercase = self.scheduler_classes[0](**A__) scheduler.set_timesteps(A__) lowercase = scheduler.timesteps[0] lowercase = scheduler.timesteps[1] lowercase = self.dummy_sample lowercase = 0.1 * sample lowercase = scheduler.step(A__ ,A__ ,A__).prev_sample lowercase = scheduler.step(A__ ,A__ ,A__).prev_sample self.assertEqual(output_a.shape ,sample.shape) self.assertEqual(output_a.shape ,output_a.shape) def A__ ( self): for timesteps in [1_0, 5_0, 1_0_0, 1_0_0_0]: self.check_over_configs(num_train_timesteps=A__) def A__ ( self): for clip_denoised in [True, False]: self.check_over_configs(clip_denoised=A__) def A__ ( self): lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config() lowercase = scheduler_class(**A__) lowercase = 1 scheduler.set_timesteps(A__) lowercase = scheduler.timesteps lowercase = torch.manual_seed(0) lowercase = self.dummy_model() lowercase = self.dummy_sample_deter * scheduler.init_noise_sigma for i, t in enumerate(A__): # 1. scale model input lowercase = scheduler.scale_model_input(A__ ,A__) # 2. predict noise residual lowercase = model(A__ ,A__) # 3. predict previous sample x_t-1 lowercase = scheduler.step(A__ ,A__ ,A__ ,generator=A__).prev_sample lowercase = pred_prev_sample lowercase = torch.sum(torch.abs(A__)) lowercase = torch.mean(torch.abs(A__)) assert abs(result_sum.item() - 192.7614) < 1E-2 assert abs(result_mean.item() - 0.2510) < 1E-3 def A__ ( self): lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config() lowercase = scheduler_class(**A__) lowercase = [1_0_6, 0] scheduler.set_timesteps(timesteps=A__) lowercase = scheduler.timesteps lowercase = torch.manual_seed(0) lowercase = self.dummy_model() lowercase = self.dummy_sample_deter * scheduler.init_noise_sigma for t in timesteps: # 1. scale model input lowercase = scheduler.scale_model_input(A__ ,A__) # 2. predict noise residual lowercase = model(A__ ,A__) # 3. predict previous sample x_t-1 lowercase = scheduler.step(A__ ,A__ ,A__ ,generator=A__).prev_sample lowercase = pred_prev_sample lowercase = torch.sum(torch.abs(A__)) lowercase = torch.mean(torch.abs(A__)) assert abs(result_sum.item() - 347.6357) < 1E-2 assert abs(result_mean.item() - 0.4527) < 1E-3 def A__ ( self): lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config() lowercase = scheduler_class(**A__) lowercase = [3_9, 3_0, 1_2, 1_5, 0] with self.assertRaises(A__ ,msg='''`timesteps` must be in descending order.'''): scheduler.set_timesteps(timesteps=A__) def A__ ( self): lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config() lowercase = scheduler_class(**A__) lowercase = [3_9, 3_0, 1_2, 1, 0] lowercase = len(A__) with self.assertRaises(A__ ,msg='''Can only pass one of `num_inference_steps` or `timesteps`.'''): scheduler.set_timesteps(num_inference_steps=A__ ,timesteps=A__) def A__ ( self): lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config() lowercase = scheduler_class(**A__) lowercase = [scheduler.config.num_train_timesteps] with self.assertRaises( A__ ,msg='''`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}''' ,): scheduler.set_timesteps(timesteps=A__)
633
import random import unittest import numpy as np import transformers from transformers import is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax if is_flax_available(): import os import jax.numpy as jnp from jax import jit from transformers import AutoTokenizer, FlaxAutoModelForCausalLM from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model lowercase__ :Optional[Any] = "0.12" # assumed parallelism: 8 if is_torch_available(): import torch def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None ): '''simple docstring''' if rng is None: lowercase = random.Random() lowercase = 1 for dim in shape: total_dims *= dim lowercase = [] for _ in range(lowerCAmelCase__ ): values.append(rng.randint(0 , vocab_size - 1 ) ) lowercase = np.array(lowerCAmelCase__ , dtype=jnp.intaa ).reshape(lowerCAmelCase__ ) return output def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=None ): '''simple docstring''' lowercase = ids_tensor(lowerCAmelCase__ , vocab_size=2 , rng=lowerCAmelCase__ ) # make sure that at least one token is attended to for each batch lowercase = 1 return attn_mask @require_flax class lowercase : lowercase_ : Any =None lowercase_ : List[str] =() def A__ ( self): lowercase , lowercase = self.model_tester.prepare_config_and_inputs_for_common() # cut to half length & take max batch_size 3 lowercase = 2 lowercase = inputs['''input_ids'''].shape[-1] // 2 lowercase = inputs['''input_ids'''][:max_batch_size, :sequence_length] lowercase = jnp.ones_like(A__) lowercase = attention_mask[:max_batch_size, :sequence_length] # generate max 5 tokens lowercase = input_ids.shape[-1] + 5 if config.eos_token_id is not None and config.pad_token_id is None: # hack to allow generate for models such as GPT2 as is done in `generate()` lowercase = config.eos_token_id return config, input_ids, attention_mask, max_length @is_pt_flax_cross_test def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = False lowercase = max_length lowercase = 0 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model_class.__name__[4:] # Skip the "Flax" at the beginning lowercase = getattr(A__ ,A__) lowercase = pt_model_class(A__).eval() lowercase = load_flax_weights_in_pytorch_model(A__ ,flax_model.params) lowercase = flax_model.generate(A__).sequences lowercase = pt_model.generate(torch.tensor(A__ ,dtype=torch.long)) if flax_generation_outputs.shape[-1] > pt_generation_outputs.shape[-1]: lowercase = flax_generation_outputs[:, : pt_generation_outputs.shape[-1]] self.assertListEqual(pt_generation_outputs.numpy().tolist() ,flax_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = False lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = True lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = False lowercase = max_length lowercase = 2 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = False lowercase = max_length lowercase = 2 lowercase = 2 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[0] ,input_ids.shape[0] * config.num_return_sequences) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = True lowercase = max_length lowercase = 0.8 lowercase = 1_0 lowercase = 0.3 lowercase = 1 lowercase = 8 lowercase = 9 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = max_length lowercase = 1 lowercase = 8 lowercase = 9 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = max_length lowercase = 2 lowercase = 1 lowercase = 8 lowercase = 9 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() # pad attention mask on the left lowercase = attention_mask.at[(0, 0)].set(0) lowercase = False lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__ ,attention_mask=A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__ ,attention_mask=A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() # pad attention mask on the left lowercase = attention_mask.at[(0, 0)].set(0) lowercase = True lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__ ,attention_mask=A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__ ,attention_mask=A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() # pad attention mask on the left lowercase = attention_mask.at[(0, 0)].set(0) lowercase = 2 lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__ ,attention_mask=A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__ ,attention_mask=A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) @require_flax class lowercase ( unittest.TestCase ): def A__ ( self): lowercase = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-bert''') lowercase = FlaxAutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-bert-flax-only''') lowercase = '''Hello world''' lowercase = tokenizer(A__ ,return_tensors='''np''').input_ids # typos are quickly detected (the correct argument is `do_sample`) with self.assertRaisesRegex(A__ ,'''do_samples'''): model.generate(A__ ,do_samples=A__) # arbitrary arguments that will not be used anywhere are also not accepted with self.assertRaisesRegex(A__ ,'''foo'''): lowercase = {'''foo''': '''bar'''} model.generate(A__ ,**A__)
633
1
from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES lowercase__ :Optional[int] = logging.get_logger(__name__) lowercase__ :Dict = OrderedDict( [ # Base model mapping ("albert", "FlaxAlbertModel"), ("bart", "FlaxBartModel"), ("beit", "FlaxBeitModel"), ("bert", "FlaxBertModel"), ("big_bird", "FlaxBigBirdModel"), ("blenderbot", "FlaxBlenderbotModel"), ("blenderbot-small", "FlaxBlenderbotSmallModel"), ("clip", "FlaxCLIPModel"), ("distilbert", "FlaxDistilBertModel"), ("electra", "FlaxElectraModel"), ("gpt-sw3", "FlaxGPT2Model"), ("gpt2", "FlaxGPT2Model"), ("gpt_neo", "FlaxGPTNeoModel"), ("gptj", "FlaxGPTJModel"), ("longt5", "FlaxLongT5Model"), ("marian", "FlaxMarianModel"), ("mbart", "FlaxMBartModel"), ("mt5", "FlaxMT5Model"), ("opt", "FlaxOPTModel"), ("pegasus", "FlaxPegasusModel"), ("regnet", "FlaxRegNetModel"), ("resnet", "FlaxResNetModel"), ("roberta", "FlaxRobertaModel"), ("roberta-prelayernorm", "FlaxRobertaPreLayerNormModel"), ("roformer", "FlaxRoFormerModel"), ("t5", "FlaxT5Model"), ("vision-text-dual-encoder", "FlaxVisionTextDualEncoderModel"), ("vit", "FlaxViTModel"), ("wav2vec2", "FlaxWav2Vec2Model"), ("whisper", "FlaxWhisperModel"), ("xglm", "FlaxXGLMModel"), ("xlm-roberta", "FlaxXLMRobertaModel"), ] ) lowercase__ :str = OrderedDict( [ # Model for pre-training mapping ("albert", "FlaxAlbertForPreTraining"), ("bart", "FlaxBartForConditionalGeneration"), ("bert", "FlaxBertForPreTraining"), ("big_bird", "FlaxBigBirdForPreTraining"), ("electra", "FlaxElectraForPreTraining"), ("longt5", "FlaxLongT5ForConditionalGeneration"), ("mbart", "FlaxMBartForConditionalGeneration"), ("mt5", "FlaxMT5ForConditionalGeneration"), ("roberta", "FlaxRobertaForMaskedLM"), ("roberta-prelayernorm", "FlaxRobertaPreLayerNormForMaskedLM"), ("roformer", "FlaxRoFormerForMaskedLM"), ("t5", "FlaxT5ForConditionalGeneration"), ("wav2vec2", "FlaxWav2Vec2ForPreTraining"), ("whisper", "FlaxWhisperForConditionalGeneration"), ("xlm-roberta", "FlaxXLMRobertaForMaskedLM"), ] ) lowercase__ :int = OrderedDict( [ # Model for Masked LM mapping ("albert", "FlaxAlbertForMaskedLM"), ("bart", "FlaxBartForConditionalGeneration"), ("bert", "FlaxBertForMaskedLM"), ("big_bird", "FlaxBigBirdForMaskedLM"), ("distilbert", "FlaxDistilBertForMaskedLM"), ("electra", "FlaxElectraForMaskedLM"), ("mbart", "FlaxMBartForConditionalGeneration"), ("roberta", "FlaxRobertaForMaskedLM"), ("roberta-prelayernorm", "FlaxRobertaPreLayerNormForMaskedLM"), ("roformer", "FlaxRoFormerForMaskedLM"), ("xlm-roberta", "FlaxXLMRobertaForMaskedLM"), ] ) lowercase__ :Tuple = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ("bart", "FlaxBartForConditionalGeneration"), ("blenderbot", "FlaxBlenderbotForConditionalGeneration"), ("blenderbot-small", "FlaxBlenderbotSmallForConditionalGeneration"), ("encoder-decoder", "FlaxEncoderDecoderModel"), ("longt5", "FlaxLongT5ForConditionalGeneration"), ("marian", "FlaxMarianMTModel"), ("mbart", "FlaxMBartForConditionalGeneration"), ("mt5", "FlaxMT5ForConditionalGeneration"), ("pegasus", "FlaxPegasusForConditionalGeneration"), ("t5", "FlaxT5ForConditionalGeneration"), ] ) lowercase__ :Any = OrderedDict( [ # Model for Image-classsification ("beit", "FlaxBeitForImageClassification"), ("regnet", "FlaxRegNetForImageClassification"), ("resnet", "FlaxResNetForImageClassification"), ("vit", "FlaxViTForImageClassification"), ] ) lowercase__ :int = OrderedDict( [ ("vision-encoder-decoder", "FlaxVisionEncoderDecoderModel"), ] ) lowercase__ :Optional[Any] = OrderedDict( [ # Model for Causal LM mapping ("bart", "FlaxBartForCausalLM"), ("bert", "FlaxBertForCausalLM"), ("big_bird", "FlaxBigBirdForCausalLM"), ("electra", "FlaxElectraForCausalLM"), ("gpt-sw3", "FlaxGPT2LMHeadModel"), ("gpt2", "FlaxGPT2LMHeadModel"), ("gpt_neo", "FlaxGPTNeoForCausalLM"), ("gptj", "FlaxGPTJForCausalLM"), ("opt", "FlaxOPTForCausalLM"), ("roberta", "FlaxRobertaForCausalLM"), ("roberta-prelayernorm", "FlaxRobertaPreLayerNormForCausalLM"), ("xglm", "FlaxXGLMForCausalLM"), ("xlm-roberta", "FlaxXLMRobertaForCausalLM"), ] ) lowercase__ :Union[str, Any] = OrderedDict( [ # Model for Sequence Classification mapping ("albert", "FlaxAlbertForSequenceClassification"), ("bart", "FlaxBartForSequenceClassification"), ("bert", "FlaxBertForSequenceClassification"), ("big_bird", "FlaxBigBirdForSequenceClassification"), ("distilbert", "FlaxDistilBertForSequenceClassification"), ("electra", "FlaxElectraForSequenceClassification"), ("mbart", "FlaxMBartForSequenceClassification"), ("roberta", "FlaxRobertaForSequenceClassification"), ("roberta-prelayernorm", "FlaxRobertaPreLayerNormForSequenceClassification"), ("roformer", "FlaxRoFormerForSequenceClassification"), ("xlm-roberta", "FlaxXLMRobertaForSequenceClassification"), ] ) lowercase__ :List[Any] = OrderedDict( [ # Model for Question Answering mapping ("albert", "FlaxAlbertForQuestionAnswering"), ("bart", "FlaxBartForQuestionAnswering"), ("bert", "FlaxBertForQuestionAnswering"), ("big_bird", "FlaxBigBirdForQuestionAnswering"), ("distilbert", "FlaxDistilBertForQuestionAnswering"), ("electra", "FlaxElectraForQuestionAnswering"), ("mbart", "FlaxMBartForQuestionAnswering"), ("roberta", "FlaxRobertaForQuestionAnswering"), ("roberta-prelayernorm", "FlaxRobertaPreLayerNormForQuestionAnswering"), ("roformer", "FlaxRoFormerForQuestionAnswering"), ("xlm-roberta", "FlaxXLMRobertaForQuestionAnswering"), ] ) lowercase__ :str = OrderedDict( [ # Model for Token Classification mapping ("albert", "FlaxAlbertForTokenClassification"), ("bert", "FlaxBertForTokenClassification"), ("big_bird", "FlaxBigBirdForTokenClassification"), ("distilbert", "FlaxDistilBertForTokenClassification"), ("electra", "FlaxElectraForTokenClassification"), ("roberta", "FlaxRobertaForTokenClassification"), ("roberta-prelayernorm", "FlaxRobertaPreLayerNormForTokenClassification"), ("roformer", "FlaxRoFormerForTokenClassification"), ("xlm-roberta", "FlaxXLMRobertaForTokenClassification"), ] ) lowercase__ :Optional[Any] = OrderedDict( [ # Model for Multiple Choice mapping ("albert", "FlaxAlbertForMultipleChoice"), ("bert", "FlaxBertForMultipleChoice"), ("big_bird", "FlaxBigBirdForMultipleChoice"), ("distilbert", "FlaxDistilBertForMultipleChoice"), ("electra", "FlaxElectraForMultipleChoice"), ("roberta", "FlaxRobertaForMultipleChoice"), ("roberta-prelayernorm", "FlaxRobertaPreLayerNormForMultipleChoice"), ("roformer", "FlaxRoFormerForMultipleChoice"), ("xlm-roberta", "FlaxXLMRobertaForMultipleChoice"), ] ) lowercase__ :Optional[Any] = OrderedDict( [ ("bert", "FlaxBertForNextSentencePrediction"), ] ) lowercase__ :Dict = OrderedDict( [ ("speech-encoder-decoder", "FlaxSpeechEncoderDecoderModel"), ("whisper", "FlaxWhisperForConditionalGeneration"), ] ) lowercase__ :Dict = OrderedDict( [ ("whisper", "FlaxWhisperForAudioClassification"), ] ) lowercase__ :int = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) lowercase__ :Optional[int] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) lowercase__ :int = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) lowercase__ :Optional[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) lowercase__ :Optional[int] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) lowercase__ :Optional[Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) lowercase__ :str = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) lowercase__ :Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) lowercase__ :Any = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) lowercase__ :int = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) lowercase__ :List[str] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) lowercase__ :List[str] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) lowercase__ :Optional[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) lowercase__ :Optional[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class lowercase ( _BaseAutoModelClass ): lowercase_ : str =FLAX_MODEL_MAPPING lowercase__ :Dict = auto_class_update(FlaxAutoModel) class lowercase ( _BaseAutoModelClass ): lowercase_ : List[str] =FLAX_MODEL_FOR_PRETRAINING_MAPPING lowercase__ :Union[str, Any] = auto_class_update(FlaxAutoModelForPreTraining, head_doc="pretraining") class lowercase ( _BaseAutoModelClass ): lowercase_ : Union[str, Any] =FLAX_MODEL_FOR_CAUSAL_LM_MAPPING lowercase__ :Any = auto_class_update(FlaxAutoModelForCausalLM, head_doc="causal language modeling") class lowercase ( _BaseAutoModelClass ): lowercase_ : Union[str, Any] =FLAX_MODEL_FOR_MASKED_LM_MAPPING lowercase__ :Dict = auto_class_update(FlaxAutoModelForMaskedLM, head_doc="masked language modeling") class lowercase ( _BaseAutoModelClass ): lowercase_ : Union[str, Any] =FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING lowercase__ :List[str] = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc="sequence-to-sequence language modeling", checkpoint_for_example="t5-base" ) class lowercase ( _BaseAutoModelClass ): lowercase_ : Tuple =FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING lowercase__ :Dict = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc="sequence classification" ) class lowercase ( _BaseAutoModelClass ): lowercase_ : Tuple =FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING lowercase__ :Tuple = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc="question answering") class lowercase ( _BaseAutoModelClass ): lowercase_ : Dict =FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING lowercase__ :Optional[Any] = auto_class_update( FlaxAutoModelForTokenClassification, head_doc="token classification" ) class lowercase ( _BaseAutoModelClass ): lowercase_ : List[str] =FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING lowercase__ :Tuple = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc="multiple choice") class lowercase ( _BaseAutoModelClass ): lowercase_ : Any =FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING lowercase__ :Optional[int] = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc="next sentence prediction" ) class lowercase ( _BaseAutoModelClass ): lowercase_ : Optional[int] =FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING lowercase__ :Any = auto_class_update( FlaxAutoModelForImageClassification, head_doc="image classification" ) class lowercase ( _BaseAutoModelClass ): lowercase_ : Any =FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING lowercase__ :List[Any] = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc="vision-to-text modeling") class lowercase ( _BaseAutoModelClass ): lowercase_ : int =FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING lowercase__ :Optional[int] = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc="sequence-to-sequence speech-to-text modeling" )
633
import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : List[str] =(UniPCMultistepScheduler,) lowercase_ : Tuple =(('''num_inference_steps''', 25),) def A__ ( self ,**A__): lowercase = { '''num_train_timesteps''': 1_0_0_0, '''beta_start''': 0.0001, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', '''solver_order''': 2, '''solver_type''': '''bh2''', } config.update(**A__) return config def A__ ( self ,A__=0 ,**A__): lowercase = dict(self.forward_default_kwargs) lowercase = kwargs.pop('''num_inference_steps''' ,A__) lowercase = self.dummy_sample lowercase = 0.1 * sample lowercase = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) scheduler.set_timesteps(A__) # copy over dummy past residuals lowercase = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A__) lowercase = scheduler_class.from_pretrained(A__) new_scheduler.set_timesteps(A__) # copy over dummy past residuals lowercase = dummy_past_residuals[: new_scheduler.config.solver_order] lowercase , lowercase = sample, sample for t in range(A__ ,time_step + scheduler.config.solver_order + 1): lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = new_scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1E-5, "Scheduler outputs are not identical" def A__ ( self ,A__=0 ,**A__): lowercase = dict(self.forward_default_kwargs) lowercase = kwargs.pop('''num_inference_steps''' ,A__) lowercase = self.dummy_sample lowercase = 0.1 * sample lowercase = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config() lowercase = scheduler_class(**A__) scheduler.set_timesteps(A__) # copy over dummy past residuals (must be after setting timesteps) lowercase = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A__) lowercase = scheduler_class.from_pretrained(A__) # copy over dummy past residuals new_scheduler.set_timesteps(A__) # copy over dummy past residual (must be after setting timesteps) lowercase = dummy_past_residuals[: new_scheduler.config.solver_order] lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = new_scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1E-5, "Scheduler outputs are not identical" def A__ ( self ,A__=None ,**A__): if scheduler is None: lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) lowercase = 1_0 lowercase = self.dummy_model() lowercase = self.dummy_sample_deter scheduler.set_timesteps(A__) for i, t in enumerate(scheduler.timesteps): lowercase = model(A__ ,A__) lowercase = scheduler.step(A__ ,A__ ,A__).prev_sample return sample def A__ ( self): lowercase = dict(self.forward_default_kwargs) lowercase = kwargs.pop('''num_inference_steps''' ,A__) for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config() lowercase = scheduler_class(**A__) lowercase = self.dummy_sample lowercase = 0.1 * sample if num_inference_steps is not None and hasattr(A__ ,'''set_timesteps'''): scheduler.set_timesteps(A__) elif num_inference_steps is not None and not hasattr(A__ ,'''set_timesteps'''): lowercase = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) lowercase = [residual + 0.2, residual + 0.15, residual + 0.10] lowercase = dummy_past_residuals[: scheduler.config.solver_order] lowercase = scheduler.timesteps[5] lowercase = scheduler.timesteps[6] lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample self.assertEqual(output_a.shape ,sample.shape) self.assertEqual(output_a.shape ,output_a.shape) def A__ ( self): # make sure that iterating over schedulers with same config names gives same results # for defaults lowercase = UniPCMultistepScheduler(**self.get_scheduler_config()) lowercase = self.full_loop(scheduler=A__) lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.2464) < 1E-3 lowercase = DPMSolverSinglestepScheduler.from_config(scheduler.config) lowercase = DEISMultistepScheduler.from_config(scheduler.config) lowercase = DPMSolverMultistepScheduler.from_config(scheduler.config) lowercase = UniPCMultistepScheduler.from_config(scheduler.config) lowercase = self.full_loop(scheduler=A__) lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.2464) < 1E-3 def A__ ( self): for timesteps in [2_5, 5_0, 1_0_0, 9_9_9, 1_0_0_0]: self.check_over_configs(num_train_timesteps=A__) def A__ ( self): self.check_over_configs(thresholding=A__) for order in [1, 2, 3]: for solver_type in ["bh1", "bh2"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=A__ ,prediction_type=A__ ,sample_max_value=A__ ,solver_order=A__ ,solver_type=A__ ,) def A__ ( self): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=A__) def A__ ( self): for solver_type in ["bh1", "bh2"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=A__ ,solver_type=A__ ,prediction_type=A__ ,) lowercase = self.full_loop( solver_order=A__ ,solver_type=A__ ,prediction_type=A__ ,) assert not torch.isnan(A__).any(), "Samples have nan numbers" def A__ ( self): self.check_over_configs(lower_order_final=A__) self.check_over_configs(lower_order_final=A__) def A__ ( self): for num_inference_steps in [1, 2, 3, 5, 1_0, 5_0, 1_0_0, 9_9_9, 1_0_0_0]: self.check_over_forward(num_inference_steps=A__ ,time_step=0) def A__ ( self): lowercase = self.full_loop() lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.2464) < 1E-3 def A__ ( self): lowercase = self.full_loop(prediction_type='''v_prediction''') lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.1014) < 1E-3 def A__ ( self): lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config(thresholding=A__ ,dynamic_thresholding_ratio=0) lowercase = scheduler_class(**A__) lowercase = 1_0 lowercase = self.dummy_model() lowercase = self.dummy_sample_deter.half() scheduler.set_timesteps(A__) for i, t in enumerate(scheduler.timesteps): lowercase = model(A__ ,A__) lowercase = scheduler.step(A__ ,A__ ,A__).prev_sample assert sample.dtype == torch.floataa def A__ ( self ,**A__): for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) scheduler.set_timesteps(scheduler.config.num_train_timesteps) assert len(scheduler.timesteps.unique()) == scheduler.num_inference_steps
633
1
def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = set() # edges = list of graph's edges lowercase = get_edges(lowerCAmelCase__ ) # While there are still elements in edges list, take an arbitrary edge # (from_node, to_node) and add his extremity to chosen_vertices and then # remove all arcs adjacent to the from_node and to_node while edges: lowercase , lowercase = edges.pop() chosen_vertices.add(lowerCAmelCase__ ) chosen_vertices.add(lowerCAmelCase__ ) for edge in edges.copy(): if from_node in edge or to_node in edge: edges.discard(lowerCAmelCase__ ) return chosen_vertices def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = set() for from_node, to_nodes in graph.items(): for to_node in to_nodes: edges.add((from_node, to_node) ) return edges if __name__ == "__main__": import doctest doctest.testmod() # graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} # print(f"Matching vertex cover:\n{matching_min_vertex_cover(graph)}")
633
import argparse from collections import defaultdict import yaml lowercase__ :Optional[int] = "docs/source/en/_toctree.yml" def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = defaultdict(lowerCAmelCase__ ) for doc in model_doc: counts[doc["local"]] += 1 lowercase = [key for key, value in counts.items() if value > 1] lowercase = [] for duplicate_key in duplicates: lowercase = list({doc['''title'''] for doc in model_doc if doc['''local'''] == duplicate_key} ) if len(lowerCAmelCase__ ) > 1: raise ValueError( f'{duplicate_key} is present several times in the documentation table of content at ' '''`docs/source/en/_toctree.yml` with different *Title* values. Choose one of those and remove the ''' '''others.''' ) # Only add this once new_doc.append({'''local''': duplicate_key, '''title''': titles[0]} ) # Add none duplicate-keys new_doc.extend([doc for doc in model_doc if counts[doc['''local''']] == 1] ) # Sort return sorted(lowerCAmelCase__ , key=lambda lowerCAmelCase__ : s["title"].lower() ) def UpperCamelCase ( lowerCAmelCase__=False ): '''simple docstring''' with open(lowerCAmelCase__ , encoding='''utf-8''' ) as f: lowercase = yaml.safe_load(f.read() ) # Get to the API doc lowercase = 0 while content[api_idx]["title"] != "API": api_idx += 1 lowercase = content[api_idx]['''sections'''] # Then to the model doc lowercase = 0 while api_doc[model_idx]["title"] != "Models": model_idx += 1 lowercase = api_doc[model_idx]['''sections'''] lowercase = [(idx, section) for idx, section in enumerate(lowerCAmelCase__ ) if '''sections''' in section] lowercase = False for idx, modality_doc in modalities_docs: lowercase = modality_doc['''sections'''] lowercase = clean_model_doc_toc(lowerCAmelCase__ ) if old_modality_doc != new_modality_doc: lowercase = True if overwrite: lowercase = new_modality_doc if diff: if overwrite: lowercase = model_doc lowercase = api_doc with open(lowerCAmelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(yaml.dump(lowerCAmelCase__ , allow_unicode=lowerCAmelCase__ ) ) else: raise ValueError( '''The model doc part of the table of content is not properly sorted, run `make style` to fix this.''' ) if __name__ == "__main__": lowercase__ :Any = argparse.ArgumentParser() parser.add_argument("--fix_and_overwrite", action="store_true", help="Whether to fix inconsistencies.") lowercase__ :int = parser.parse_args() check_model_doc(args.fix_and_overwrite)
633
1
from ...utils import ( OptionalDependencyNotAvailable, is_flax_available, is_torch_available, is_transformers_available, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .multicontrolnet import MultiControlNetModel from .pipeline_controlnet import StableDiffusionControlNetPipeline from .pipeline_controlnet_imgaimg import StableDiffusionControlNetImgaImgPipeline from .pipeline_controlnet_inpaint import StableDiffusionControlNetInpaintPipeline if is_transformers_available() and is_flax_available(): from .pipeline_flax_controlnet import FlaxStableDiffusionControlNetPipeline
633
from __future__ import annotations from decimal import Decimal from math import * # noqa: F403 from sympy import diff def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 10**-10 ): '''simple docstring''' lowercase = a while True: lowercase = Decimal(lowerCAmelCase__ ) - ( Decimal(eval(lowerCAmelCase__ ) ) / Decimal(eval(str(diff(lowerCAmelCase__ ) ) ) ) # noqa: S307 ) # This number dictates the accuracy of the answer if abs(eval(lowerCAmelCase__ ) ) < precision: # noqa: S307 return float(lowerCAmelCase__ ) # Let's Execute if __name__ == "__main__": # Find root of trigonometric function # Find value of pi print(F'The root of sin(x) = 0 is {newton_raphson("sin(x)", 2)}') # Find root of polynomial print(F'The root of x**2 - 5*x + 2 = 0 is {newton_raphson("x**2 - 5*x + 2", 0.4)}') # Find Square Root of 5 print(F'The root of log(x) - 1 = 0 is {newton_raphson("log(x) - 1", 2)}') # Exponential Roots print(F'The root of exp(x) - 1 = 0 is {newton_raphson("exp(x) - 1", 0)}')
633
1
import inspect import os import unittest import torch import accelerate from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_multi_gpu from accelerate.utils import patch_environment class lowercase ( unittest.TestCase ): def A__ ( self): lowercase = inspect.getfile(accelerate.test_utils) lowercase = os.path.sep.join(mod_file.split(os.path.sep)[:-1] + ['''scripts''', '''test_script.py''']) lowercase = os.path.sep.join( mod_file.split(os.path.sep)[:-1] + ['''scripts''', '''test_distributed_data_loop.py''']) lowercase = os.path.sep.join(mod_file.split(os.path.sep)[:-1] + ['''scripts''', '''test_ops.py''']) @require_multi_gpu def A__ ( self): print(f'Found {torch.cuda.device_count()} devices.') lowercase = ['''torchrun''', f'--nproc_per_node={torch.cuda.device_count()}', self.test_file_path] with patch_environment(omp_num_threads=1): execute_subprocess_async(A__ ,env=os.environ.copy()) @require_multi_gpu def A__ ( self): print(f'Found {torch.cuda.device_count()} devices.') lowercase = ['''torchrun''', f'--nproc_per_node={torch.cuda.device_count()}', self.operation_file_path] print(f'Command: {cmd}') with patch_environment(omp_num_threads=1): execute_subprocess_async(A__ ,env=os.environ.copy()) @require_multi_gpu def A__ ( self): lowercase = ['''torchrun''', f'--nproc_per_node={torch.cuda.device_count()}', inspect.getfile(self.__class__)] with patch_environment(omp_num_threads=1): execute_subprocess_async(A__ ,env=os.environ.copy()) @require_multi_gpu def A__ ( self): print(f'Found {torch.cuda.device_count()} devices, using 2 devices only') lowercase = ['''torchrun''', f'--nproc_per_node={torch.cuda.device_count()}', self.data_loop_file_path] with patch_environment(omp_num_threads=1 ,cuda_visible_devices='''0,1'''): execute_subprocess_async(A__ ,env=os.environ.copy()) if __name__ == "__main__": lowercase__ :int = Accelerator() lowercase__ :Optional[int] = (accelerator.state.process_index + 2, 10) lowercase__ :List[str] = torch.randint(0, 10, shape).to(accelerator.device) lowercase__ :Any = "" lowercase__ :List[str] = accelerator.pad_across_processes(tensor) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." if not torch.equal(tensora[: accelerator.state.process_index + 2], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[accelerator.state.process_index + 2 :] == 0): error_msg += "Padding was not done with the right value (0)." lowercase__ :Dict = accelerator.pad_across_processes(tensor, pad_first=True) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." lowercase__ :Tuple = accelerator.state.num_processes - accelerator.state.process_index - 1 if not torch.equal(tensora[index:], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[:index] == 0): error_msg += "Padding was not done with the right value (0)." # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
633
from pathlib import Path import numpy as np from PIL import Image def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase , lowercase = rgb[:, :, 0], rgb[:, :, 1], rgb[:, :, 2] return 0.29_89 * r + 0.58_70 * g + 0.11_40 * b def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return (gray > 127) & (gray <= 255) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = np.zeros_like(lowerCAmelCase__ ) lowercase = np.zeros( (image.shape[0] + kernel.shape[0] - 1, image.shape[1] + kernel.shape[1] - 1) ) # Copy image to padded image lowercase = image # Iterate over image & apply kernel for x in range(image.shape[1] ): for y in range(image.shape[0] ): lowercase = ( kernel * image_padded[y : y + kernel.shape[0], x : x + kernel.shape[1]] ).sum() lowercase = int(summation > 0 ) return output if __name__ == "__main__": # read original image lowercase__ :str = Path(__file__).resolve().parent / "image_data" / "lena.jpg" lowercase__ :List[str] = np.array(Image.open(lena_path)) # kernel to be applied lowercase__ :Union[str, Any] = np.array([[0, 1, 0], [1, 1, 1], [0, 1, 0]]) lowercase__ :Optional[int] = dilation(gray_to_binary(rgb_to_gray(lena)), structuring_element) # Save the output image lowercase__ :str = Image.fromarray(output).convert("RGB") pil_img.save("result_dilation.png")
633
1
lowercase__ :List[str] = "0.21.0" from .accelerator import Accelerator from .big_modeling import ( cpu_offload, cpu_offload_with_hook, disk_offload, dispatch_model, init_empty_weights, init_on_device, load_checkpoint_and_dispatch, ) from .data_loader import skip_first_batches from .launchers import debug_launcher, notebook_launcher from .state import PartialState from .utils import ( DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, FullyShardedDataParallelPlugin, GradScalerKwargs, InitProcessGroupKwargs, find_executable_batch_size, infer_auto_device_map, is_rich_available, load_checkpoint_in_model, synchronize_rng_states, ) if is_rich_available(): from .utils import rich
633
import argparse import torch from torch import nn from transformers import MaMaaaConfig, MaMaaaForConditionalGeneration def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = [ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''decoder.output_projection.weight''', '''_float_tensor''', '''encoder.embed_positions._float_tensor''', '''decoder.embed_positions._float_tensor''', ] for k in ignore_keys: state_dict.pop(lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase = emb.weight.shape lowercase = nn.Linear(lowerCAmelCase__ , lowerCAmelCase__ , bias=lowerCAmelCase__ ) lowercase = emb.weight.data return lin_layer def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = torch.load(lowerCAmelCase__ , map_location='''cpu''' ) lowercase = mam_aaa['''args'''] or mam_aaa['''cfg''']['''model'''] lowercase = mam_aaa['''model'''] remove_ignore_keys_(lowerCAmelCase__ ) lowercase = state_dict['''encoder.embed_tokens.weight'''].shape[0] lowercase = MaMaaaConfig( vocab_size=lowerCAmelCase__ , max_position_embeddings=1024 , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , encoder_layerdrop=args.encoder_layerdrop , decoder_layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='''relu''' , ) lowercase = state_dict['''decoder.embed_tokens.weight'''] lowercase = MaMaaaForConditionalGeneration(lowerCAmelCase__ ) model.model.load_state_dict(lowerCAmelCase__ , strict=lowerCAmelCase__ ) lowercase = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": lowercase__ :Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument("fairseq_path", type=str, help="path to a model.pt on local filesystem.") parser.add_argument("pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") lowercase__ :Tuple = parser.parse_args() lowercase__ :int = convert_fairseq_mamaaa_checkpoint_from_disk(args.fairseq_pathß) model.save_pretrained(args.pytorch_dump_folder_path)
633
1
def UpperCamelCase ( lowerCAmelCase__ = 100_0000 ): '''simple docstring''' lowercase = [i - 1 for i in range(limit + 1 )] for i in range(2 , limit + 1 ): if phi[i] == i - 1: for j in range(2 * i , limit + 1 , lowerCAmelCase__ ): phi[j] -= phi[j] // i return sum(phi[2 : limit + 1] ) if __name__ == "__main__": print(solution())
633
from __future__ import annotations from random import random class lowercase : def __init__( self ,A__ = None): lowercase = value lowercase = random() lowercase = None lowercase = None def __repr__( self): from pprint import pformat if self.left is None and self.right is None: return f'\'{self.value}: {self.prior:.5}\'' else: return pformat( {f'{self.value}: {self.prior:.5}': (self.left, self.right)} ,indent=1) def __str__( self): lowercase = str(self.value) + ''' ''' lowercase = str(self.left or '''''') lowercase = str(self.right or '''''') return value + left + right def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' if root is None: # None tree is split into 2 Nones return None, None elif root.value is None: return None, None else: if value < root.value: lowercase , lowercase = split(root.left , lowerCAmelCase__ ) return left, root else: lowercase , lowercase = split(root.right , lowerCAmelCase__ ) return root, right def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' if (not left) or (not right): # If one node is None, return the other return left or right elif left.prior < right.prior: lowercase = merge(left.right , lowerCAmelCase__ ) return left else: lowercase = merge(lowerCAmelCase__ , right.left ) return right def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = Node(lowerCAmelCase__ ) lowercase , lowercase = split(lowerCAmelCase__ , lowerCAmelCase__ ) return merge(merge(lowerCAmelCase__ , lowerCAmelCase__ ) , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase = split(lowerCAmelCase__ , value - 1 ) lowercase , lowercase = split(lowerCAmelCase__ , lowerCAmelCase__ ) return merge(lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if not root: # None return else: inorder(root.left ) print(root.value , end=''',''' ) inorder(root.right ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' for arg in args.split(): if arg[0] == "+": lowercase = insert(lowerCAmelCase__ , int(arg[1:] ) ) elif arg[0] == "-": lowercase = erase(lowerCAmelCase__ , int(arg[1:] ) ) else: print('''Unknown command''' ) return root def UpperCamelCase ( ): '''simple docstring''' lowercase = None print( '''enter numbers to create a tree, + value to add value into treap, ''' '''- value to erase all nodes with value. \'q\' to quit. ''' ) lowercase = input() while args != "q": lowercase = interact_treap(lowerCAmelCase__ , lowerCAmelCase__ ) print(lowerCAmelCase__ ) lowercase = input() print('''good by!''' ) if __name__ == "__main__": import doctest doctest.testmod() main()
633
1
import os import shutil from pathlib import Path from typing import Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ..utils import ONNX_EXTERNAL_WEIGHTS_NAME, ONNX_WEIGHTS_NAME, is_onnx_available, logging if is_onnx_available(): import onnxruntime as ort lowercase__ :List[Any] = logging.get_logger(__name__) lowercase__ :Dict = { "tensor(bool)": np.bool_, "tensor(int8)": np.inta, "tensor(uint8)": np.uinta, "tensor(int16)": np.intaa, "tensor(uint16)": np.uintaa, "tensor(int32)": np.intaa, "tensor(uint32)": np.uintaa, "tensor(int64)": np.intaa, "tensor(uint64)": np.uintaa, "tensor(float16)": np.floataa, "tensor(float)": np.floataa, "tensor(double)": np.floataa, } class lowercase : def __init__( self ,A__=None ,**A__): logger.info('''`diffusers.OnnxRuntimeModel` is experimental and might change in the future.''') lowercase = model lowercase = kwargs.get('''model_save_dir''' ,A__) lowercase = kwargs.get('''latest_model_name''' ,A__) def __call__( self ,**A__): lowercase = {k: np.array(A__) for k, v in kwargs.items()} return self.model.run(A__ ,A__) @staticmethod def A__ ( A__ ,A__=None ,A__=None): if provider is None: logger.info('''No onnxruntime provider specified, using CPUExecutionProvider''') lowercase = '''CPUExecutionProvider''' return ort.InferenceSession(A__ ,providers=[provider] ,sess_options=A__) def A__ ( self ,A__ ,A__ = None ,**A__): lowercase = file_name if file_name is not None else ONNX_WEIGHTS_NAME lowercase = self.model_save_dir.joinpath(self.latest_model_name) lowercase = Path(A__).joinpath(A__) try: shutil.copyfile(A__ ,A__) except shutil.SameFileError: pass # copy external weights (for models >2GB) lowercase = self.model_save_dir.joinpath(A__) if src_path.exists(): lowercase = Path(A__).joinpath(A__) try: shutil.copyfile(A__ ,A__) except shutil.SameFileError: pass def A__ ( self ,A__ ,**A__ ,): if os.path.isfile(A__): logger.error(f'Provided path ({save_directory}) should be a directory, not a file') return os.makedirs(A__ ,exist_ok=A__) # saving model weights/files self._save_pretrained(A__ ,**A__) @classmethod def A__ ( cls ,A__ ,A__ = None ,A__ = None ,A__ = False ,A__ = None ,A__ = None ,A__ = None ,A__ = None ,**A__ ,): lowercase = file_name if file_name is not None else ONNX_WEIGHTS_NAME # load model from local directory if os.path.isdir(A__): lowercase = OnnxRuntimeModel.load_model( os.path.join(A__ ,A__) ,provider=A__ ,sess_options=A__) lowercase = Path(A__) # load model from hub else: # download model lowercase = hf_hub_download( repo_id=A__ ,filename=A__ ,use_auth_token=A__ ,revision=A__ ,cache_dir=A__ ,force_download=A__ ,) lowercase = Path(A__).parent lowercase = Path(A__).name lowercase = OnnxRuntimeModel.load_model(A__ ,provider=A__ ,sess_options=A__) return cls(model=A__ ,**A__) @classmethod def A__ ( cls ,A__ ,A__ = True ,A__ = None ,A__ = None ,**A__ ,): lowercase = None if len(str(A__).split('''@''')) == 2: lowercase , lowercase = model_id.split('''@''') return cls._from_pretrained( model_id=A__ ,revision=A__ ,cache_dir=A__ ,force_download=A__ ,use_auth_token=A__ ,**A__ ,)
633
def UpperCamelCase ( lowerCAmelCase__ = 1000 ): '''simple docstring''' lowercase = -1 lowercase = 0 for a in range(1 , n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N eliminating c lowercase = (n * n - 2 * a * n) // (2 * n - 2 * a) lowercase = n - a - b if c * c == (a * a + b * b): lowercase = a * b * c if candidate >= product: lowercase = candidate return product if __name__ == "__main__": print(F'{solution() = }')
633
1
from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ :Dict = logging.get_logger(__name__) lowercase__ :int = {} class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Union[str, Any] ='''llama''' lowercase_ : List[Any] =['''past_key_values'''] def __init__( self ,A__=3_2_0_0_0 ,A__=4_0_9_6 ,A__=1_1_0_0_8 ,A__=3_2 ,A__=3_2 ,A__=None ,A__="silu" ,A__=2_0_4_8 ,A__=0.02 ,A__=1E-6 ,A__=True ,A__=0 ,A__=1 ,A__=2 ,A__=1 ,A__=False ,A__=None ,**A__ ,): lowercase = vocab_size lowercase = max_position_embeddings lowercase = hidden_size lowercase = intermediate_size lowercase = num_hidden_layers lowercase = num_attention_heads # for backward compatibility if num_key_value_heads is None: lowercase = num_attention_heads lowercase = num_key_value_heads lowercase = hidden_act lowercase = initializer_range lowercase = rms_norm_eps lowercase = pretraining_tp lowercase = use_cache lowercase = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=A__ ,bos_token_id=A__ ,eos_token_id=A__ ,tie_word_embeddings=A__ ,**A__ ,) def A__ ( self): if self.rope_scaling is None: return if not isinstance(self.rope_scaling ,A__) or len(self.rope_scaling) != 2: raise ValueError( '''`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, ''' f'got {self.rope_scaling}') lowercase = self.rope_scaling.get('''type''' ,A__) lowercase = self.rope_scaling.get('''factor''' ,A__) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f'`rope_scaling`\'s name field must be one of [\'linear\', \'dynamic\'], got {rope_scaling_type}') if rope_scaling_factor is None or not isinstance(A__ ,A__) or rope_scaling_factor <= 1.0: raise ValueError(f'`rope_scaling`\'s factor field must be an float > 1, got {rope_scaling_factor}')
633
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ :Tuple = { "configuration_instructblip": [ "INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "InstructBlipConfig", "InstructBlipQFormerConfig", "InstructBlipVisionConfig", ], "processing_instructblip": ["InstructBlipProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :List[str] = [ "INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "InstructBlipQFormerModel", "InstructBlipPreTrainedModel", "InstructBlipForConditionalGeneration", "InstructBlipVisionModel", ] if TYPE_CHECKING: from .configuration_instructblip import ( INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, InstructBlipConfig, InstructBlipQFormerConfig, InstructBlipVisionConfig, ) from .processing_instructblip import InstructBlipProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_instructblip import ( INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST, InstructBlipForConditionalGeneration, InstructBlipPreTrainedModel, InstructBlipQFormerModel, InstructBlipVisionModel, ) else: import sys lowercase__ :List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
633
1
import numpy as np import torch from torch.utils.data import Dataset from utils import logger class lowercase ( SCREAMING_SNAKE_CASE__ ): def __init__( self ,A__ ,A__): lowercase = params lowercase = np.array(A__) lowercase = np.array([len(A__) for t in data]) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__( self ,A__): return (self.token_ids[index], self.lengths[index]) def __len__( self): return len(self.lengths) def A__ ( self): assert len(self.token_ids) == len(self.lengths) assert all(self.lengths[i] == len(self.token_ids[i]) for i in range(len(self.lengths))) def A__ ( self): lowercase = self.params.max_model_input_size lowercase = self.lengths > max_len logger.info(f'Splitting {sum(A__)} too long sequences.') def divide_chunks(A__ ,A__): return [l[i : i + n] for i in range(0 ,len(A__) ,A__)] lowercase = [] lowercase = [] if self.params.mlm: lowercase , lowercase = self.params.special_tok_ids['''cls_token'''], self.params.special_tok_ids['''sep_token'''] else: lowercase , lowercase = self.params.special_tok_ids['''bos_token'''], self.params.special_tok_ids['''eos_token'''] for seq_, len_ in zip(self.token_ids ,self.lengths): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_) new_lengths.append(len_) else: lowercase = [] for sub_s in divide_chunks(seq_ ,max_len - 2): if sub_s[0] != cls_id: lowercase = np.insert(A__ ,0 ,A__) if sub_s[-1] != sep_id: lowercase = np.insert(A__ ,len(A__) ,A__) assert len(A__) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(A__) new_tok_ids.extend(A__) new_lengths.extend([len(A__) for l in sub_seqs]) lowercase = np.array(A__) lowercase = np.array(A__) def A__ ( self): lowercase = len(self) lowercase = self.lengths > 1_1 lowercase = self.token_ids[indices] lowercase = self.lengths[indices] lowercase = len(self) logger.info(f'Remove {init_size - new_size} too short (<=11 tokens) sequences.') def A__ ( self): if "unk_token" not in self.params.special_tok_ids: return else: lowercase = self.params.special_tok_ids['''unk_token'''] lowercase = len(self) lowercase = np.array([np.count_nonzero(a == unk_token_id) for a in self.token_ids]) lowercase = (unk_occs / self.lengths) < 0.5 lowercase = self.token_ids[indices] lowercase = self.lengths[indices] lowercase = len(self) logger.info(f'Remove {init_size - new_size} sequences with a high level of unknown tokens (50%).') def A__ ( self): if not self.params.is_master: return logger.info(f'{len(self)} sequences') # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def A__ ( self ,A__): lowercase = [t[0] for t in batch] lowercase = [t[1] for t in batch] assert len(A__) == len(A__) # Max for paddings lowercase = max(A__) # Pad token ids if self.params.mlm: lowercase = self.params.special_tok_ids['''pad_token'''] else: lowercase = self.params.special_tok_ids['''unk_token'''] lowercase = [list(t.astype(A__)) + [pad_idx] * (max_seq_len_ - len(A__)) for t in token_ids] assert len(tk_) == len(A__) assert all(len(A__) == max_seq_len_ for t in tk_) lowercase = torch.tensor(tk_) # (bs, max_seq_len_) lowercase = torch.tensor(A__) # (bs) return tk_t, lg_t
633
from numpy import exp, pi, sqrt def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ = 0.0 , lowerCAmelCase__ = 1.0 ): '''simple docstring''' return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
633
1
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ :Dict = logging.get_logger(__name__) lowercase__ :Optional[int] = { "microsoft/wavlm-base": "https://huggingface.co/microsoft/wavlm-base/resolve/main/config.json", # See all WavLM models at https://huggingface.co/models?filter=wavlm } class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Optional[int] ='''wavlm''' def __init__( self ,A__=3_2 ,A__=7_6_8 ,A__=1_2 ,A__=1_2 ,A__=3_0_7_2 ,A__="gelu" ,A__=0.1 ,A__=0.1 ,A__=0.1 ,A__=0.0 ,A__=0.1 ,A__=0.1 ,A__=0.02 ,A__=1E-5 ,A__="group" ,A__="gelu" ,A__=(5_1_2, 5_1_2, 5_1_2, 5_1_2, 5_1_2, 5_1_2, 5_1_2) ,A__=(5, 2, 2, 2, 2, 2, 2) ,A__=(1_0, 3, 3, 3, 3, 2, 2) ,A__=False ,A__=1_2_8 ,A__=1_6 ,A__=3_2_0 ,A__=8_0_0 ,A__=False ,A__=True ,A__=0.05 ,A__=1_0 ,A__=2 ,A__=0.0 ,A__=1_0 ,A__=3_2_0 ,A__=2 ,A__=0.1 ,A__=1_0_0 ,A__=2_5_6 ,A__=2_5_6 ,A__=0.1 ,A__="mean" ,A__=False ,A__=False ,A__=2_5_6 ,A__=(5_1_2, 5_1_2, 5_1_2, 5_1_2, 1_5_0_0) ,A__=(5, 3, 3, 1, 1) ,A__=(1, 2, 3, 1, 1) ,A__=5_1_2 ,A__=8_0 ,A__=0 ,A__=1 ,A__=2 ,A__=False ,A__=3 ,A__=2 ,A__=3 ,A__=None ,**A__ ,): super().__init__(**A__ ,pad_token_id=A__ ,bos_token_id=A__ ,eos_token_id=A__) lowercase = hidden_size lowercase = feat_extract_norm lowercase = feat_extract_activation lowercase = list(A__) lowercase = list(A__) lowercase = list(A__) lowercase = conv_bias lowercase = num_buckets lowercase = max_bucket_distance lowercase = num_conv_pos_embeddings lowercase = num_conv_pos_embedding_groups lowercase = len(self.conv_dim) lowercase = num_hidden_layers lowercase = intermediate_size lowercase = hidden_act lowercase = num_attention_heads lowercase = hidden_dropout lowercase = attention_dropout lowercase = activation_dropout lowercase = feat_proj_dropout lowercase = final_dropout lowercase = layerdrop lowercase = layer_norm_eps lowercase = initializer_range lowercase = num_ctc_classes lowercase = vocab_size lowercase = do_stable_layer_norm lowercase = use_weighted_layer_sum lowercase = classifier_proj_size if ( (len(self.conv_stride) != self.num_feat_extract_layers) or (len(self.conv_kernel) != self.num_feat_extract_layers) or (len(self.conv_dim) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==''' ''' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =''' f' {len(self.conv_dim)}`, `len(config.conv_stride) = {len(self.conv_stride)}`,' f' `len(config.conv_kernel) = {len(self.conv_kernel)}`.') # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 lowercase = apply_spec_augment lowercase = mask_time_prob lowercase = mask_time_length lowercase = mask_time_min_masks lowercase = mask_feature_prob lowercase = mask_feature_length # parameters for pretraining with codevector quantized representations lowercase = num_codevectors_per_group lowercase = num_codevector_groups lowercase = contrastive_logits_temperature lowercase = num_negatives lowercase = codevector_dim lowercase = proj_codevector_dim lowercase = diversity_loss_weight # ctc loss lowercase = ctc_loss_reduction lowercase = ctc_zero_infinity # adapter lowercase = add_adapter lowercase = adapter_kernel_size lowercase = adapter_stride lowercase = num_adapter_layers lowercase = output_hidden_size or hidden_size # SequenceClassification-specific parameter. Feel free to ignore for other classes. lowercase = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. lowercase = list(A__) lowercase = list(A__) lowercase = list(A__) lowercase = xvector_output_dim @property def A__ ( self): return functools.reduce(operator.mul ,self.conv_stride ,1)
633
import asyncio import os import shutil import subprocess import sys import tempfile import unittest from distutils.util import strtobool from functools import partial from pathlib import Path from typing import List, Union from unittest import mock import torch from ..state import AcceleratorState, PartialState from ..utils import ( gather, is_bnb_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_mps_available, is_safetensors_available, is_tensorboard_available, is_torch_version, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=False ): '''simple docstring''' try: lowercase = os.environ[key] except KeyError: # KEY isn't set, default to `default`. lowercase = default else: # KEY is set, convert it to True or False. try: lowercase = strtobool(lowerCAmelCase__ ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(f'If set, {key} must be yes or no.' ) return _value lowercase__ :Dict = parse_flag_from_env("RUN_SLOW", default=False) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skip('''Test was skipped''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(_run_slow_tests , '''test is slow''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(not torch.cuda.is_available() , '''test requires only a CPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.cuda.is_available() , '''test requires a GPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_xpu_available() , '''test requires a XPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_mps_available() , '''test requires a `mps` backend support in `torch`''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless( is_transformers_available() and is_datasets_available() , '''test requires the Hugging Face suite''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_bnb_available() , '''test requires the bitsandbytes library''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_tpu_available() , '''test requires TPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.cuda.device_count() == 1 , '''test requires a GPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.xpu.device_count() == 1 , '''test requires a XPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.cuda.device_count() > 1 , '''test requires multiple GPUs''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.xpu.device_count() > 1 , '''test requires multiple XPUs''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_safetensors_available() , '''test requires safetensors''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_deepspeed_available() , '''test requires DeepSpeed''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_torch_version('''>=''' , '''1.12.0''' ) , '''test requires torch version >= 1.12.0''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__=None , lowerCAmelCase__=None ): '''simple docstring''' if test_case is None: return partial(lowerCAmelCase__ , version=lowerCAmelCase__ ) return unittest.skipUnless(is_torch_version('''>=''' , lowerCAmelCase__ ) , f'test requires torch version >= {version}' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_tensorboard_available() , '''test requires Tensorboard''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_wandb_available() , '''test requires wandb''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_comet_ml_available() , '''test requires comet_ml''' )(lowerCAmelCase__ ) lowercase__ :Dict = ( any([is_wandb_available(), is_tensorboard_available()]) and not is_comet_ml_available() ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless( _atleast_one_tracker_available , '''test requires at least one tracker to be available and for `comet_ml` to not be installed''' , )(lowerCAmelCase__ ) class lowercase ( unittest.TestCase ): lowercase_ : int =True @classmethod def A__ ( cls): lowercase = tempfile.mkdtemp() @classmethod def A__ ( cls): if os.path.exists(cls.tmpdir): shutil.rmtree(cls.tmpdir) def A__ ( self): if self.clear_on_setup: for path in Path(self.tmpdir).glob('''**/*'''): if path.is_file(): path.unlink() elif path.is_dir(): shutil.rmtree(A__) class lowercase ( unittest.TestCase ): def A__ ( self): super().tearDown() # Reset the state of the AcceleratorState singleton. AcceleratorState._reset_state() PartialState._reset_state() class lowercase ( unittest.TestCase ): def A__ ( self ,A__): lowercase = mocks if isinstance(A__ ,(tuple, list)) else [mocks] for m in self.mocks: m.start() self.addCleanup(m.stop) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = AcceleratorState() lowercase = tensor[None].clone().to(state.device ) lowercase = gather(lowerCAmelCase__ ).cpu() lowercase = tensor[0].cpu() for i in range(tensors.shape[0] ): if not torch.equal(tensors[i] , lowerCAmelCase__ ): return False return True class lowercase : def __init__( self ,A__ ,A__ ,A__): lowercase = returncode lowercase = stdout lowercase = stderr async def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' while True: lowercase = await stream.readline() if line: callback(lowerCAmelCase__ ) else: break async def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=False , lowerCAmelCase__=False ): '''simple docstring''' if echo: print('''\nRunning: ''' , ''' '''.join(lowerCAmelCase__ ) ) lowercase = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=lowerCAmelCase__ , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=lowerCAmelCase__ , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) lowercase = [] lowercase = [] def tee(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__="" ): lowercase = line.decode('''utf-8''' ).rstrip() sink.append(lowerCAmelCase__ ) if not quiet: print(lowerCAmelCase__ , lowerCAmelCase__ , file=lowerCAmelCase__ ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ asyncio.create_task(_read_stream(p.stdout , lambda lowerCAmelCase__ : tee(lowerCAmelCase__ , lowerCAmelCase__ , sys.stdout , label='''stdout:''' ) ) ), asyncio.create_task(_read_stream(p.stderr , lambda lowerCAmelCase__ : tee(lowerCAmelCase__ , lowerCAmelCase__ , sys.stderr , label='''stderr:''' ) ) ), ] , timeout=lowerCAmelCase__ , ) return _RunOutput(await p.wait() , lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=180 , lowerCAmelCase__=False , lowerCAmelCase__=True ): '''simple docstring''' lowercase = asyncio.get_event_loop() lowercase = loop.run_until_complete( _stream_subprocess(lowerCAmelCase__ , env=lowerCAmelCase__ , stdin=lowerCAmelCase__ , timeout=lowerCAmelCase__ , quiet=lowerCAmelCase__ , echo=lowerCAmelCase__ ) ) lowercase = ''' '''.join(lowerCAmelCase__ ) if result.returncode > 0: lowercase = '''\n'''.join(result.stderr ) raise RuntimeError( f'\'{cmd_str}\' failed with returncode {result.returncode}\n\n' f'The combined stderr from workers follows:\n{stderr}' ) return result class lowercase ( SCREAMING_SNAKE_CASE__ ): pass def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=False ): '''simple docstring''' try: lowercase = subprocess.check_output(lowerCAmelCase__ , stderr=subprocess.STDOUT ) if return_stdout: if hasattr(lowerCAmelCase__ , '''decode''' ): lowercase = output.decode('''utf-8''' ) return output except subprocess.CalledProcessError as e: raise SubprocessCallException( f'Command `{" ".join(lowerCAmelCase__ )}` failed with the following error:\n\n{e.output.decode()}' ) from e
633
1
def UpperCamelCase ( lowerCAmelCase__ = 1000 ): '''simple docstring''' return sum(2 * a * ((a - 1) // 2) for a in range(3 , n + 1 ) ) if __name__ == "__main__": print(solution())
633
import math import time from typing import Dict, List, Optional from torch.utils.data import Dataset from transformers import SeqaSeqTrainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class lowercase ( SCREAMING_SNAKE_CASE__ ): def __init__( self ,*A__ ,A__=None ,A__=None ,**A__): super().__init__(*A__ ,**A__) lowercase = eval_examples lowercase = post_process_function def A__ ( self ,A__ = None ,A__=None ,A__ = None ,A__ = "eval" ,**A__ ,): lowercase = gen_kwargs.copy() lowercase = ( gen_kwargs['''max_length'''] if gen_kwargs.get('''max_length''') is not None else self.args.generation_max_length ) lowercase = ( gen_kwargs['''num_beams'''] if gen_kwargs.get('''num_beams''') is not None else self.args.generation_num_beams ) lowercase = gen_kwargs lowercase = self.eval_dataset if eval_dataset is None else eval_dataset lowercase = self.get_eval_dataloader(A__) lowercase = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. lowercase = self.compute_metrics lowercase = None lowercase = time.time() lowercase = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: lowercase = eval_loop( A__ ,description='''Evaluation''' ,prediction_loss_only=True if compute_metrics is None else None ,ignore_keys=A__ ,metric_key_prefix=A__ ,) finally: lowercase = compute_metrics lowercase = self.args.eval_batch_size * self.args.world_size if f'{metric_key_prefix}_jit_compilation_time' in output.metrics: start_time += output.metrics[f'{metric_key_prefix}_jit_compilation_time'] output.metrics.update( speed_metrics( A__ ,A__ ,num_samples=output.num_samples ,num_steps=math.ceil(output.num_samples / total_batch_size) ,)) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default lowercase = self.post_process_function(A__ ,A__ ,A__) lowercase = self.compute_metrics(A__) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys()): if not key.startswith(f'{metric_key_prefix}_'): lowercase = metrics.pop(A__) metrics.update(output.metrics) else: lowercase = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(A__) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report()) lowercase = self.callback_handler.on_evaluate(self.args ,self.state ,self.control ,A__) return metrics def A__ ( self ,A__ ,A__ ,A__=None ,A__ = "test" ,**A__): lowercase = gen_kwargs.copy() lowercase = self.get_test_dataloader(A__) # Temporarily disable metric computation, we will do it in the loop here. lowercase = self.compute_metrics lowercase = None lowercase = time.time() lowercase = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: lowercase = eval_loop( A__ ,description='''Prediction''' ,prediction_loss_only=True if compute_metrics is None else None ,ignore_keys=A__ ,metric_key_prefix=A__ ,) finally: lowercase = compute_metrics lowercase = self.args.eval_batch_size * self.args.world_size if f'{metric_key_prefix}_jit_compilation_time' in output.metrics: start_time += output.metrics[f'{metric_key_prefix}_jit_compilation_time'] output.metrics.update( speed_metrics( A__ ,A__ ,num_samples=output.num_samples ,num_steps=math.ceil(output.num_samples / total_batch_size) ,)) if self.post_process_function is None or self.compute_metrics is None: return output lowercase = self.post_process_function(A__ ,A__ ,A__ ,'''predict''') lowercase = self.compute_metrics(A__) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys()): if not key.startswith(f'{metric_key_prefix}_'): lowercase = metrics.pop(A__) metrics.update(output.metrics) return PredictionOutput(predictions=predictions.predictions ,label_ids=predictions.label_ids ,metrics=A__)
633
1
from __future__ import annotations from statistics import mean def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = [0] * no_of_processes lowercase = [0] * no_of_processes # Initialize remaining_time to waiting_time. for i in range(lowerCAmelCase__ ): lowercase = burst_time[i] lowercase = [] lowercase = 0 lowercase = 0 # When processes are not completed, # A process whose arrival time has passed \ # and has remaining execution time is put into the ready_process. # The shortest process in the ready_process, target_process is executed. while completed != no_of_processes: lowercase = [] lowercase = -1 for i in range(lowerCAmelCase__ ): if (arrival_time[i] <= total_time) and (remaining_time[i] > 0): ready_process.append(lowerCAmelCase__ ) if len(lowerCAmelCase__ ) > 0: lowercase = ready_process[0] for i in ready_process: if remaining_time[i] < remaining_time[target_process]: lowercase = i total_time += burst_time[target_process] completed += 1 lowercase = 0 lowercase = ( total_time - arrival_time[target_process] - burst_time[target_process] ) else: total_time += 1 return waiting_time def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = [0] * no_of_processes for i in range(lowerCAmelCase__ ): lowercase = burst_time[i] + waiting_time[i] return turn_around_time if __name__ == "__main__": print("[TEST CASE 01]") lowercase__ :Tuple = 4 lowercase__ :Dict = [2, 5, 3, 7] lowercase__ :int = [0, 0, 0, 0] lowercase__ :Dict = calculate_waitingtime(arrival_time, burst_time, no_of_processes) lowercase__ :int = calculate_turnaroundtime( burst_time, no_of_processes, waiting_time ) # Printing the Result print("PID\tBurst Time\tArrival Time\tWaiting Time\tTurnaround Time") for i, process_id in enumerate(list(range(1, 5))): print( F'{process_id}\t{burst_time[i]}\t\t\t{arrival_time[i]}\t\t\t\t' F'{waiting_time[i]}\t\t\t\t{turn_around_time[i]}' ) print(F'\nAverage waiting time = {mean(waiting_time):.5f}') print(F'Average turnaround time = {mean(turn_around_time):.5f}')
633
# Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import subprocess from packaging.version import Version, parse from accelerate.commands.config.config_args import default_config_file, load_config_from_file lowercase__ :Any = "Run commands across TPU VMs for initial setup before running `accelerate launch`." def UpperCamelCase ( lowerCAmelCase__=None ): '''simple docstring''' if subparsers is not None: lowercase = subparsers.add_parser('''tpu-config''' , description=_description ) else: lowercase = argparse.ArgumentParser('''Accelerate tpu-config command''' , description=_description ) # Core arguments lowercase = parser.add_argument_group( '''Config Arguments''' , '''Arguments that can be configured through `accelerate config`.''' ) config_args.add_argument( '''--config_file''' , type=lowerCAmelCase__ , default=lowerCAmelCase__ , help='''Path to the config file to use for accelerate.''' , ) config_args.add_argument( '''--tpu_name''' , default=lowerCAmelCase__ , help='''The name of the TPU to use. If not specified, will use the TPU specified in the config file.''' , ) config_args.add_argument( '''--tpu_zone''' , default=lowerCAmelCase__ , help='''The zone of the TPU to use. If not specified, will use the zone specified in the config file.''' , ) lowercase = parser.add_argument_group('''TPU Arguments''' , '''Arguments for options ran inside the TPU.''' ) pod_args.add_argument( '''--use_alpha''' , action='''store_true''' , help='''Whether to use `gcloud alpha` when running the TPU training script instead of `gcloud`.''' , ) pod_args.add_argument( '''--command_file''' , default=lowerCAmelCase__ , help='''The path to the file containing the commands to run on the pod on startup.''' , ) pod_args.add_argument( '''--command''' , action='''append''' , nargs='''+''' , help='''A command to run on the pod. Can be passed multiple times.''' , ) pod_args.add_argument( '''--install_accelerate''' , action='''store_true''' , help='''Whether to install accelerate on the pod. Defaults to False.''' , ) pod_args.add_argument( '''--accelerate_version''' , default='''latest''' , help='''The version of accelerate to install on the pod. If not specified, will use the latest pypi version. Specify \'dev\' to install from GitHub.''' , ) pod_args.add_argument( '''--debug''' , action='''store_true''' , help='''If set, will print the command that would be run instead of running it.''' ) if subparsers is not None: parser.set_defaults(func=lowerCAmelCase__ ) return parser def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = None # Get the default from the config file if it exists. if args.config_file is not None or os.path.isfile(lowerCAmelCase__ ): lowercase = load_config_from_file(args.config_file ) if not args.command_file and defaults.command_file is not None and not args.command: lowercase = defaults.command_file if not args.command and defaults.commands is not None: lowercase = defaults.commands if not args.tpu_name: lowercase = defaults.tpu_name if not args.tpu_zone: lowercase = defaults.tpu_zone if args.accelerate_version == "dev": lowercase = '''git+https://github.com/huggingface/accelerate.git''' elif args.accelerate_version == "latest": lowercase = '''accelerate -U''' elif isinstance(parse(args.accelerate_version ) , lowerCAmelCase__ ): lowercase = f'accelerate=={args.accelerate_version}' if not args.command_file and not args.command: raise ValueError('''You must specify either a command file or a command to run on the pod.''' ) if args.command_file: with open(args.command_file , '''r''' ) as f: lowercase = [f.read().splitlines()] # To turn list of lists into list of strings if isinstance(args.command[0] , lowerCAmelCase__ ): lowercase = [line for cmd in args.command for line in cmd] # Default to the shared folder and install accelerate lowercase = ['''cd /usr/share'''] if args.install_accelerate: new_cmd += [f'pip install {args.accelerate_version}'] new_cmd += args.command lowercase = '''; '''.join(lowerCAmelCase__ ) # Then send it to gcloud # Eventually try to use google-api-core to do this instead of subprocess lowercase = ['''gcloud'''] if args.use_alpha: cmd += ["alpha"] cmd += [ "compute", "tpus", "tpu-vm", "ssh", args.tpu_name, "--zone", args.tpu_zone, "--command", args.command, "--worker", "all", ] if args.debug: print(f'Running {" ".join(lowerCAmelCase__ )}' ) return subprocess.run(lowerCAmelCase__ ) print('''Successfully setup pod.''' ) def UpperCamelCase ( ): '''simple docstring''' lowercase = tpu_command_parser() lowercase = parser.parse_args() tpu_command_launcher(lowerCAmelCase__ )
633
1
# this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won't be considered # since the output of this script is fed into Makefile commands it doesn't print a newline after the results import re import subprocess import sys lowercase__ :List[Any] = subprocess.check_output("git merge-base main HEAD".split()).decode("utf-8") lowercase__ :str = ( subprocess.check_output(F'git diff --diff-filter=d --name-only {fork_point_sha}'.split()).decode("utf-8").split() ) lowercase__ :int = "|".join(sys.argv[1:]) lowercase__ :Union[str, Any] = re.compile(rF'^({joined_dirs}).*?\.py$') lowercase__ :str = [x for x in modified_files if regex.match(x)] print(" ".join(relevant_modified_files), end="")
633
import json import multiprocessing as mp import re from collections import defaultdict from functools import partial from typing import Dict, List, Optional, Set, Tuple, Type from datasets import Dataset from datasketch import MinHash, MinHashLSH from dpu_utils.utils.iterators import ThreadedIterator from tqdm import tqdm lowercase__ :Optional[Any] = re.compile("[^A-Za-z_0-9]") # parameters used in DuplicationIndex lowercase__ :List[str] = 10 lowercase__ :Tuple = 256 def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if len(lowerCAmelCase__ ) < MIN_NUM_TOKENS: return None lowercase = MinHash(num_perm=lowerCAmelCase__ ) for token in set(lowerCAmelCase__ ): min_hash.update(token.encode() ) return min_hash def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return {t for t in NON_ALPHA.split(lowerCAmelCase__ ) if len(t.strip() ) > 0} class lowercase : def __init__( self ,*, A__ = 0.85 ,): lowercase = duplication_jaccard_threshold lowercase = NUM_PERM lowercase = MinHashLSH(threshold=self._duplication_jaccard_threshold ,num_perm=self._num_perm) lowercase = defaultdict(A__) def A__ ( self ,A__ ,A__): lowercase = self._index.query(A__) if code_key in self._index.keys: print(f'Duplicate key {code_key}') return self._index.insert(A__ ,A__) if len(A__) > 0: for base_duplicate in close_duplicates: if base_duplicate in self._duplicate_clusters: self._duplicate_clusters[base_duplicate].add(A__) break else: self._duplicate_clusters[close_duplicates[0]].add(A__) def A__ ( self): lowercase = [] for base, duplicates in self._duplicate_clusters.items(): lowercase = [base] + list(A__) # reformat the cluster to be a list of dict lowercase = [{'''base_index''': el[0], '''repo_name''': el[1], '''path''': el[2]} for el in cluster] duplicate_clusters.append(A__) return duplicate_clusters def A__ ( self ,A__): lowercase = self.get_duplicate_clusters() with open(A__ ,'''w''') as f: json.dump(A__ ,A__) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase = element lowercase = get_min_hash([t for t in NON_ALPHA.split(data['''content'''] ) if len(t.strip() ) > 0] ) if min_hash is not None: return (index, data["repo_name"], data["path"]), min_hash def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' with mp.Pool() as pool: for data in pool.imap_unordered( _compute_min_hash , ThreadedIterator(lowerCAmelCase__ , max_queue_size=1_0000 ) , chunksize=100 , ): if data is not None: yield data def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = DuplicationIndex(duplication_jaccard_threshold=lowerCAmelCase__ ) for filename, min_hash in tqdm(ThreadedIterator(minhash_iter(enumerate(lowerCAmelCase__ ) ) , max_queue_size=100 ) ): di.add(lowerCAmelCase__ , lowerCAmelCase__ ) # Returns a List[Cluster] where Cluster is List[str] with the filenames. return di.get_duplicate_clusters() def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = get_tokens(lowerCAmelCase__ ) lowercase = get_tokens(lowerCAmelCase__ ) return len(tokensa & tokensa ) / len(tokensa | tokensa ) lowercase__ :List[Any] = None def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = [] for elementa in cluster: lowercase = _shared_dataset[elementa['''base_index''']]['''content'''] for elementa in extremes: lowercase = _shared_dataset[elementa['''base_index''']]['''content'''] if jaccard_similarity(lowerCAmelCase__ , lowerCAmelCase__ ) >= jaccard_threshold: elementa["copies"] += 1 break else: lowercase = 1 extremes.append(lowerCAmelCase__ ) return extremes def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' global _shared_dataset lowercase = dataset lowercase = [] lowercase = partial(_find_cluster_extremes_shared , jaccard_threshold=lowerCAmelCase__ ) with mp.Pool() as pool: for extremes in tqdm( pool.imap_unordered( lowerCAmelCase__ , lowerCAmelCase__ , ) , total=len(lowerCAmelCase__ ) , ): extremes_list.append(lowerCAmelCase__ ) return extremes_list def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ = 0.85 ): '''simple docstring''' lowercase = make_duplicate_clusters(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase = {x['''base_index'''] for cluster in duplicate_clusters for x in cluster} lowercase = {} lowercase = find_extremes(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) for extremes in extremes_clusters: for element in extremes: lowercase = element lowercase = duplicate_indices - set(extreme_dict.keys() ) lowercase = dataset.filter(lambda lowerCAmelCase__ , lowerCAmelCase__ : idx not in remove_indices , with_indices=lowerCAmelCase__ ) # update duplicate_clusters for cluster in duplicate_clusters: for element in cluster: lowercase = element['''base_index'''] in extreme_dict if element["is_extreme"]: lowercase = extreme_dict[element['''base_index''']]['''copies'''] print(f'Original dataset size: {len(lowerCAmelCase__ )}' ) print(f'Number of duplicate clusters: {len(lowerCAmelCase__ )}' ) print(f'Files in duplicate cluster: {len(lowerCAmelCase__ )}' ) print(f'Unique files in duplicate cluster: {len(lowerCAmelCase__ )}' ) print(f'Filtered dataset size: {len(lowerCAmelCase__ )}' ) return ds_filter, duplicate_clusters
633
1
def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return 10 - x * x def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' # Bolzano theory in order to find if there is a root between a and b if equation(lowerCAmelCase__ ) * equation(lowerCAmelCase__ ) >= 0: raise ValueError('''Wrong space!''' ) lowercase = a while (b - a) >= 0.01: # Find middle point lowercase = (a + b) / 2 # Check if middle point is root if equation(lowerCAmelCase__ ) == 0.0: break # Decide the side to repeat the steps if equation(lowerCAmelCase__ ) * equation(lowerCAmelCase__ ) < 0: lowercase = c else: lowercase = c return c if __name__ == "__main__": import doctest doctest.testmod() print(bisection(-2, 5)) print(bisection(0, 6))
633
import numpy as np import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModelWithProjection, PreTrainedModel from ...utils import logging lowercase__ :Union[str, Any] = logging.get_logger(__name__) class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Union[str, Any] =CLIPConfig lowercase_ : str =['''CLIPEncoderLayer'''] def __init__( self ,A__): super().__init__(A__) lowercase = CLIPVisionModelWithProjection(config.vision_config) lowercase = nn.Linear(config.vision_config.projection_dim ,1) lowercase = nn.Linear(config.vision_config.projection_dim ,1) @torch.no_grad() def A__ ( self ,A__ ,A__ ,A__=0.5 ,A__=0.5): lowercase = self.vision_model(A__)[0] lowercase = self.p_head(A__) lowercase = nsfw_detected.flatten() lowercase = nsfw_detected > p_threshold lowercase = nsfw_detected.tolist() if any(A__): logger.warning( '''Potential NSFW content was detected in one or more images. A black image will be returned instead.''' ''' Try again with a different prompt and/or seed.''') for idx, nsfw_detected_ in enumerate(A__): if nsfw_detected_: lowercase = np.zeros(images[idx].shape) lowercase = self.w_head(A__) lowercase = watermark_detected.flatten() lowercase = watermark_detected > w_threshold lowercase = watermark_detected.tolist() if any(A__): logger.warning( '''Potential watermarked content was detected in one or more images. A black image will be returned instead.''' ''' Try again with a different prompt and/or seed.''') for idx, watermark_detected_ in enumerate(A__): if watermark_detected_: lowercase = np.zeros(images[idx].shape) return images, nsfw_detected, watermark_detected
633
1
def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' # if the collection is empty, returns empty if collection == []: return [] # get some information about the collection lowercase = len(lowerCAmelCase__ ) lowercase = max(lowerCAmelCase__ ) lowercase = min(lowerCAmelCase__ ) # create the counting array lowercase = coll_max + 1 - coll_min lowercase = [0] * counting_arr_length # count how much a number appears in the collection for number in collection: counting_arr[number - coll_min] += 1 # sum each position with it's predecessors. now, counting_arr[i] tells # us how many elements <= i has in the collection for i in range(1 , lowerCAmelCase__ ): lowercase = counting_arr[i] + counting_arr[i - 1] # create the output collection lowercase = [0] * coll_len # place the elements in the output, respecting the original order (stable # sort) from end to begin, updating counting_arr for i in reversed(range(0 , lowerCAmelCase__ ) ): lowercase = collection[i] counting_arr[collection[i] - coll_min] -= 1 return ordered def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return "".join([chr(lowerCAmelCase__ ) for i in counting_sort([ord(lowerCAmelCase__ ) for c in string] )] ) if __name__ == "__main__": # Test string sort assert counting_sort_string("thisisthestring") == "eghhiiinrsssttt" lowercase__ :int = input("Enter numbers separated by a comma:\n").strip() lowercase__ :str = [int(item) for item in user_input.split(",")] print(counting_sort(unsorted))
633
class lowercase : def __init__( self ,A__): lowercase = val lowercase = None lowercase = None def A__ ( self ,A__): if self.val: if val < self.val: if self.left is None: lowercase = Node(A__) else: self.left.insert(A__) elif val > self.val: if self.right is None: lowercase = Node(A__) else: self.right.insert(A__) else: lowercase = val def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' # Recursive traversal if root: inorder(root.left , lowerCAmelCase__ ) res.append(root.val ) inorder(root.right , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' # Build BST if len(lowerCAmelCase__ ) == 0: return arr lowercase = Node(arr[0] ) for i in range(1 , len(lowerCAmelCase__ ) ): root.insert(arr[i] ) # Traverse BST in order. lowercase = [] inorder(lowerCAmelCase__ , lowerCAmelCase__ ) return res if __name__ == "__main__": print(tree_sort([10, 1, 3, 2, 9, 14, 13]))
633
1
def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return sum(i for i in range(1 , number // 2 + 1 ) if number % i == 0 ) == number if __name__ == "__main__": print("Program to check whether a number is a Perfect number or not...") lowercase__ :int = int(input("Enter number: ").strip()) print(F'{number} is {"" if perfect(number) else "not "}a Perfect Number.')
633
import os def UpperCamelCase ( lowerCAmelCase__ = "input.txt" ): '''simple docstring''' with open(os.path.join(os.path.dirname(lowerCAmelCase__ ) , lowerCAmelCase__ ) ) as input_file: lowercase = [ [int(lowerCAmelCase__ ) for element in line.split(''',''' )] for line in input_file.readlines() ] lowercase = len(lowerCAmelCase__ ) lowercase = len(matrix[0] ) lowercase = [[-1 for _ in range(lowerCAmelCase__ )] for _ in range(lowerCAmelCase__ )] for i in range(lowerCAmelCase__ ): lowercase = matrix[i][0] for j in range(1 , lowerCAmelCase__ ): for i in range(lowerCAmelCase__ ): lowercase = minimal_path_sums[i][j - 1] + matrix[i][j] for i in range(1 , lowerCAmelCase__ ): lowercase = min( minimal_path_sums[i][j] , minimal_path_sums[i - 1][j] + matrix[i][j] ) for i in range(rows - 2 , -1 , -1 ): lowercase = min( minimal_path_sums[i][j] , minimal_path_sums[i + 1][j] + matrix[i][j] ) return min(minimal_path_sums_row[-1] for minimal_path_sums_row in minimal_path_sums ) if __name__ == "__main__": print(F'{solution() = }')
633
1
from __future__ import annotations def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if len(lowerCAmelCase__ ) < 2: raise ValueError('''Monogons and Digons are not polygons in the Euclidean space''' ) if any(i <= 0 for i in nums ): raise ValueError('''All values must be greater than 0''' ) lowercase = nums.copy() copy_nums.sort() return copy_nums[-1] < sum(copy_nums[:-1] ) if __name__ == "__main__": import doctest doctest.testmod()
633
from __future__ import annotations def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if len(lowerCAmelCase__ ) < 2: raise ValueError('''Monogons and Digons are not polygons in the Euclidean space''' ) if any(i <= 0 for i in nums ): raise ValueError('''All values must be greater than 0''' ) lowercase = nums.copy() copy_nums.sort() return copy_nums[-1] < sum(copy_nums[:-1] ) if __name__ == "__main__": import doctest doctest.testmod()
633
1
from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake lowercase__ :Tuple = numpy.array([0, 0]) lowercase__ :Tuple = numpy.array([0.5, 0.8_660_254]) lowercase__ :Optional[Any] = numpy.array([1, 0]) lowercase__ :str = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = initial_vectors for _ in range(lowerCAmelCase__ ): lowercase = iteration_step(lowerCAmelCase__ ) return vectors def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = [] for i, start_vector in enumerate(vectors[:-1] ): lowercase = vectors[i + 1] new_vectors.append(lowerCAmelCase__ ) lowercase = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = numpy.radians(lowerCAmelCase__ ) lowercase , lowercase = numpy.cos(lowerCAmelCase__ ), numpy.sin(lowerCAmelCase__ ) lowercase = numpy.array(((c, -s), (s, c)) ) return numpy.dot(lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() lowercase , lowercase = zip(*lowerCAmelCase__ ) plt.plot(lowerCAmelCase__ , lowerCAmelCase__ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() lowercase__ :Tuple = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
633
def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' return int((input_a, input_a).count(0 ) == 0 ) def UpperCamelCase ( ): '''simple docstring''' assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_gate(1 , 0 ) == 0 assert and_gate(1 , 1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
633
1
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowercase__ :int = {"configuration_van": ["VAN_PRETRAINED_CONFIG_ARCHIVE_MAP", "VanConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :Optional[int] = [ "VAN_PRETRAINED_MODEL_ARCHIVE_LIST", "VanForImageClassification", "VanModel", "VanPreTrainedModel", ] if TYPE_CHECKING: from .configuration_van import VAN_PRETRAINED_CONFIG_ARCHIVE_MAP, VanConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_van import ( VAN_PRETRAINED_MODEL_ARCHIVE_LIST, VanForImageClassification, VanModel, VanPreTrainedModel, ) else: import sys lowercase__ :Union[str, Any] = _LazyModule(__name__, globals()["__file__"], _import_structure)
633
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowercase__ :Tuple = { "configuration_biogpt": ["BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BioGptConfig"], "tokenization_biogpt": ["BioGptTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :Union[str, Any] = [ "BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST", "BioGptForCausalLM", "BioGptForTokenClassification", "BioGptForSequenceClassification", "BioGptModel", "BioGptPreTrainedModel", ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys lowercase__ :Any = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
633
1
import collections import json import os import re from typing import TYPE_CHECKING, List, Optional, Tuple import numpy as np from ...tokenization_utils_fast import PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation lowercase__ :List[str] = logging.get_logger(__name__) lowercase__ :List[str] = {"vocab_file": "vocab.txt", "emoji_file": "emoji.json"} lowercase__ :Dict = { "vocab_file": { "abeja/gpt-neox-japanese-2.7b": "https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/vocab.txt", }, "emoji_file": { "abeja/gpt-neox-japanese-2.7b": "https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/emoji.json", }, } lowercase__ :Optional[Any] = { "abeja/gpt-neox-japanese-2.7b": 2048, } def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' with open(lowerCAmelCase__ , '''r''' , encoding='''utf-8''' ) as f: lowercase = json.loads(f.read() ) lowercase = collections.OrderedDict() lowercase = collections.OrderedDict() lowercase = collections.OrderedDict() with open(lowerCAmelCase__ , '''r''' , encoding='''utf-8''' ) as f: lowercase = f.readlines() lowercase = [[t.rstrip('''\n''' )] if (t == ''',''' or ''',''' not in t) else t.rstrip('''\n''' ).split(''',''' ) for t in token] for idx, b in enumerate(lowerCAmelCase__ ): lowercase = b lowercase = idx for wd in b: lowercase = idx return vocab, raw_vocab, ids_to_tokens, emoji class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Union[str, Any] =VOCAB_FILES_NAMES lowercase_ : Any =PRETRAINED_VOCAB_FILES_MAP lowercase_ : Optional[int] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ : Tuple =['''input_ids''', '''attention_mask'''] def __init__( self ,A__ ,A__ ,A__="<|endoftext|>" ,A__="<|endoftext|>" ,A__="<|startoftext|>" ,A__="<|endoftext|>" ,A__=False ,**A__ ,): super().__init__( unk_token=A__ ,pad_token=A__ ,bos_token=A__ ,eos_token=A__ ,do_clean_text=A__ ,**A__ ,) if not os.path.isfile(A__): raise ValueError( f'Can\'t find a vocabulary file at path \'{vocab_file}\'. To load the vocabulary from a Google pretrained' ''' model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`''') if not os.path.isfile(A__): raise ValueError( f'Can\'t find a emoji file at path \'{emoji_file}\'. To load the emoji information from a Google' ''' pretrained model use `tokenizer = GPTNeoXJapaneseokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`''') lowercase = do_clean_text lowercase , lowercase , lowercase , lowercase = load_vocab_and_emoji(A__ ,A__) lowercase = SubWordJapaneseTokenizer( vocab=self.vocab ,ids_to_tokens=self.ids_to_tokens ,emoji=self.emoji) @property def A__ ( self): # self.vocab contains support for character fluctuation unique to Japanese, and has a large number of vocab return len(self.raw_vocab) def A__ ( self): return dict(self.raw_vocab ,**self.added_tokens_encoder) def A__ ( self ,A__): return self.subword_tokenizer.tokenize(A__ ,clean=self.do_clean_text) def A__ ( self ,A__): return self.vocab.get(A__ ,self.vocab.get(self.unk_token)) def A__ ( self ,A__): return self.subword_tokenizer.convert_id_to_token(A__) def A__ ( self ,A__): lowercase = ''''''.join(A__).strip() return out_string def A__ ( self ,A__): lowercase = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(A__ ,add_special_tokens=A__) + [self.eos_token_id]) if len(A__) > self.model_max_length: lowercase = input_ids[-self.model_max_length :] return input_ids def A__ ( self ,A__ ,A__ = None): lowercase = 0 if os.path.isdir(A__): lowercase = os.path.join( A__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file''']) lowercase = os.path.join( A__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''emoji_file''']) else: lowercase = ( (filename_prefix + '''-''' if filename_prefix else '''''') + save_directory + VOCAB_FILES_NAMES['''vocab_file'''] ) lowercase = ( (filename_prefix + '''-''' if filename_prefix else '''''') + save_directory + VOCAB_FILES_NAMES['''emoji_file'''] ) with open(A__ ,'''w''' ,encoding='''utf-8''') as writer: for token_index, token in self.ids_to_tokens.items(): if index != token_index: logger.warning( f'Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.' ''' Please check that the vocabulary is not corrupted!''') lowercase = token_index writer.write(''','''.join(A__) + '''\n''') index += 1 with open(A__ ,'''w''' ,encoding='''utf-8''') as writer: json.dump(self.emoji ,A__) return vocab_file, emoji_file class lowercase ( SCREAMING_SNAKE_CASE__ ): def __init__( self ,A__ ,A__ ,A__): lowercase = vocab # same as swe lowercase = ids_to_tokens # same as bpe lowercase = emoji lowercase = np.max([len(A__) for w in self.vocab.keys()]) lowercase = re.compile(r'''(https?|ftp)(:\/\/[-_\.!~*\'()a-zA-Z0-9;\/?:\@&=\+$,%#]+)''') lowercase = re.compile(r'''[A-Za-z0-9\._+]*@[\-_0-9A-Za-z]+(\.[A-Za-z]+)*''') lowercase = re.compile(r'''[\(]{0,1}[0-9]{2,4}[\)\-\(]{0,1}[0-9]{2,4}[\)\-]{0,1}[0-9]{3,4}''') lowercase = re.compile( r'''([12]\d{3}[/\-年])*(0?[1-9]|1[0-2])[/\-月]((0?[1-9]|[12][0-9]|3[01])日?)*(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*''') lowercase = re.compile( r'''(明治|大正|昭和|平成|令和|㍾|㍽|㍼|㍻|\u32ff)\d{1,2}年(0?[1-9]|1[0-2])月(0?[1-9]|[12][0-9]|3[01])日(\d{1,2}|:|\d{1,2}時|\d{1,2}分|\(日\)|\(月\)|\(火\)|\(水\)|\(木\)|\(金\)|\(土\)|㈰|㈪|㈫|㈬|㈭|㈮|㈯)*''') lowercase = re.compile( r'''((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*億)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*万)*((0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*千)*(0|[1-9]\d*|[1-9]\d{0,2}(,\d{3})+)*(千円|万円|千万円|円|千ドル|万ドル|千万ドル|ドル|千ユーロ|万ユーロ|千万ユーロ|ユーロ)+(\(税込\)|\(税抜\)|\+tax)*''') lowercase = '''─━│┃┄┅┆┇┈┉┊┋┌┍┎┏┐┑┒┓└┕┖┗┘┙┚┛├┝┞┟┠┡┢┣┤┥┦┧┨┩┪┫┬┭┮┯┰┱┲┳┴┵┶┷┸┹┺┻┼┽┾┿╀╁╂╃╄╅╆╇╈╉╊╋╌╍╎╏═║╒╓╔╕╖╗╘╙╚╛╜╝╞╟╠╡╢╣╤╥╦╧╨╩╪╫╬╭╮╯╰╱╲╳╴╵╶╷╸╹╺╻╼╽╾╿''' lowercase = '''▀▁▂▃▄▅▆▇█▉▊▋▌▍▎▏▐░▒▓▔▕▖▗▘▙▚▛▜▝▞▟''' lowercase = str.maketrans({k: '''<BLOCK>''' for k in keisen + blocks}) def __len__( self): return len(self.ids_to_tokens) def A__ ( self ,A__): lowercase = self.content_repattera.sub('''<URL>''' ,A__) lowercase = self.content_repattera.sub('''<EMAIL>''' ,A__) lowercase = self.content_repattera.sub('''<TEL>''' ,A__) lowercase = self.content_repattera.sub('''<DATE>''' ,A__) lowercase = self.content_repattera.sub('''<DATE>''' ,A__) lowercase = self.content_repattera.sub('''<PRICE>''' ,A__) lowercase = content.translate(self.content_transa) while "<BLOCK><BLOCK>" in content: lowercase = content.replace('''<BLOCK><BLOCK>''' ,'''<BLOCK>''') return content def A__ ( self ,A__ ,A__=False): lowercase = text.replace(''' ''' ,'''<SP>''') lowercase = text.replace(''' ''' ,'''<SP>''') lowercase = text.replace('''\r\n''' ,'''<BR>''') lowercase = text.replace('''\n''' ,'''<BR>''') lowercase = text.replace('''\r''' ,'''<BR>''') lowercase = text.replace('''\t''' ,'''<TAB>''') lowercase = text.replace('''—''' ,'''ー''') lowercase = text.replace('''−''' ,'''ー''') for k, v in self.emoji["emoji"].items(): if k in text: lowercase = text.replace(A__ ,A__) if clean: lowercase = self.clean_text(A__) def check_simbol(A__): lowercase = x.encode() if len(A__) == 1 and len(A__) == 2: lowercase = (int(e[0]) << 8) + int(e[1]) if ( (c >= 0Xc2_a1 and c <= 0Xc2_bf) or (c >= 0Xc7_80 and c <= 0Xc7_83) or (c >= 0Xca_b9 and c <= 0Xcb_bf) or (c >= 0Xcc_80 and c <= 0Xcd_a2) ): return True return False def checkuae(A__): lowercase = x.encode() if len(A__) == 1 and len(A__) == 3: lowercase = (int(e[0]) << 1_6) + (int(e[1]) << 8) + int(e[2]) if c >= 0Xe2_80_80 and c <= 0Xe2_b0_7f: return True return False lowercase = 0 lowercase = [] while pos < len(A__): lowercase = min(len(A__) ,pos + self.maxlen + 1) if text[pos] == '''<''' else pos + 3 lowercase = [] # (token_id, token, pos) for e in range(A__ ,A__ ,-1): lowercase = text[pos:e] if wd in self.vocab: if wd[0] == "<" and len(A__) > 2: lowercase = [(self.vocab[wd], wd, e)] break else: candidates.append((self.vocab[wd], wd, e)) if len(A__) > 0: # the smallest token_id is adopted lowercase , lowercase , lowercase = sorted(A__ ,key=lambda A__: x[0])[0] result.append(A__) lowercase = e else: lowercase = pos + 1 lowercase = text[pos:end] if check_simbol(A__): result.append('''<KIGOU>''') elif checkuae(A__): result.append('''<U2000U2BFF>''') else: for i in wd.encode('''utf-8'''): result.append('''<|byte%d|>''' % i) lowercase = end return result def A__ ( self ,A__ ,A__="\n"): lowercase = [] lowercase = [] lowercase = self.ids_to_tokens[index][0] if word[:6] == "<|byte" and word[-2:] == "|>": byte_tokens.append(int(word[6:-2])) else: if len(A__) > 0: words.append(bytearray(A__).decode('''utf-8''' ,errors='''replace''')) lowercase = [] if word[:7] == "<|emoji" and word[-2:] == "|>": words.append(self.emoji['''emoji_inv'''][word]) elif word == "<SP>": words.append(''' ''') elif word == "<BR>": words.append(A__) elif word == "<TAB>": words.append('''\t''') elif word == "<BLOCK>": words.append('''▀''') elif word == "<KIGOU>": words.append('''ǀ''') elif word == "<U2000U2BFF>": words.append('''‖''') else: words.append(A__) if len(A__) > 0: words.append(bytearray(A__).decode('''utf-8''' ,errors='''replace''')) lowercase = ''''''.join(A__) return text
633
import logging from transformers import PretrainedConfig lowercase__ :int = logging.getLogger(__name__) lowercase__ :Dict = { "bertabs-finetuned-cnndm": "https://huggingface.co/remi/bertabs-finetuned-cnndm-extractive-abstractive-summarization/resolve/main/config.json", } class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Optional[int] ='''bertabs''' def __init__( self ,A__=3_0_5_2_2 ,A__=5_1_2 ,A__=6 ,A__=5_1_2 ,A__=8 ,A__=5_1_2 ,A__=0.2 ,A__=6 ,A__=7_6_8 ,A__=8 ,A__=2_0_4_8 ,A__=0.2 ,**A__ ,): super().__init__(**A__) lowercase = vocab_size lowercase = max_pos lowercase = enc_layers lowercase = enc_hidden_size lowercase = enc_heads lowercase = enc_ff_size lowercase = enc_dropout lowercase = dec_layers lowercase = dec_hidden_size lowercase = dec_heads lowercase = dec_ff_size lowercase = dec_dropout
633
1
def UpperCamelCase ( lowerCAmelCase__ = 100 ): '''simple docstring''' lowercase = n * (n + 1) * (2 * n + 1) / 6 lowercase = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(F'{solution() = }')
633
import warnings from diffusers import StableDiffusionInpaintPipeline as StableDiffusionInpaintPipeline # noqa F401 warnings.warn( "The `inpainting.py` script is outdated. Please use directly `from diffusers import" " StableDiffusionInpaintPipeline` instead." )
633
1
import argparse import torch from torch import nn from transformers import MaMaaaConfig, MaMaaaForConditionalGeneration def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = [ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''decoder.output_projection.weight''', '''_float_tensor''', '''encoder.embed_positions._float_tensor''', '''decoder.embed_positions._float_tensor''', ] for k in ignore_keys: state_dict.pop(lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase = emb.weight.shape lowercase = nn.Linear(lowerCAmelCase__ , lowerCAmelCase__ , bias=lowerCAmelCase__ ) lowercase = emb.weight.data return lin_layer def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = torch.load(lowerCAmelCase__ , map_location='''cpu''' ) lowercase = mam_aaa['''args'''] or mam_aaa['''cfg''']['''model'''] lowercase = mam_aaa['''model'''] remove_ignore_keys_(lowerCAmelCase__ ) lowercase = state_dict['''encoder.embed_tokens.weight'''].shape[0] lowercase = MaMaaaConfig( vocab_size=lowerCAmelCase__ , max_position_embeddings=1024 , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , encoder_layerdrop=args.encoder_layerdrop , decoder_layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='''relu''' , ) lowercase = state_dict['''decoder.embed_tokens.weight'''] lowercase = MaMaaaForConditionalGeneration(lowerCAmelCase__ ) model.model.load_state_dict(lowerCAmelCase__ , strict=lowerCAmelCase__ ) lowercase = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": lowercase__ :Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument("fairseq_path", type=str, help="path to a model.pt on local filesystem.") parser.add_argument("pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") lowercase__ :Tuple = parser.parse_args() lowercase__ :int = convert_fairseq_mamaaa_checkpoint_from_disk(args.fairseq_pathß) model.save_pretrained(args.pytorch_dump_folder_path)
633
import random import unittest import numpy as np import transformers from transformers import is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax if is_flax_available(): import os import jax.numpy as jnp from jax import jit from transformers import AutoTokenizer, FlaxAutoModelForCausalLM from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model lowercase__ :Optional[Any] = "0.12" # assumed parallelism: 8 if is_torch_available(): import torch def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None ): '''simple docstring''' if rng is None: lowercase = random.Random() lowercase = 1 for dim in shape: total_dims *= dim lowercase = [] for _ in range(lowerCAmelCase__ ): values.append(rng.randint(0 , vocab_size - 1 ) ) lowercase = np.array(lowerCAmelCase__ , dtype=jnp.intaa ).reshape(lowerCAmelCase__ ) return output def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=None ): '''simple docstring''' lowercase = ids_tensor(lowerCAmelCase__ , vocab_size=2 , rng=lowerCAmelCase__ ) # make sure that at least one token is attended to for each batch lowercase = 1 return attn_mask @require_flax class lowercase : lowercase_ : Any =None lowercase_ : List[str] =() def A__ ( self): lowercase , lowercase = self.model_tester.prepare_config_and_inputs_for_common() # cut to half length & take max batch_size 3 lowercase = 2 lowercase = inputs['''input_ids'''].shape[-1] // 2 lowercase = inputs['''input_ids'''][:max_batch_size, :sequence_length] lowercase = jnp.ones_like(A__) lowercase = attention_mask[:max_batch_size, :sequence_length] # generate max 5 tokens lowercase = input_ids.shape[-1] + 5 if config.eos_token_id is not None and config.pad_token_id is None: # hack to allow generate for models such as GPT2 as is done in `generate()` lowercase = config.eos_token_id return config, input_ids, attention_mask, max_length @is_pt_flax_cross_test def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = False lowercase = max_length lowercase = 0 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model_class.__name__[4:] # Skip the "Flax" at the beginning lowercase = getattr(A__ ,A__) lowercase = pt_model_class(A__).eval() lowercase = load_flax_weights_in_pytorch_model(A__ ,flax_model.params) lowercase = flax_model.generate(A__).sequences lowercase = pt_model.generate(torch.tensor(A__ ,dtype=torch.long)) if flax_generation_outputs.shape[-1] > pt_generation_outputs.shape[-1]: lowercase = flax_generation_outputs[:, : pt_generation_outputs.shape[-1]] self.assertListEqual(pt_generation_outputs.numpy().tolist() ,flax_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = False lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = True lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = False lowercase = max_length lowercase = 2 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = False lowercase = max_length lowercase = 2 lowercase = 2 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[0] ,input_ids.shape[0] * config.num_return_sequences) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = True lowercase = max_length lowercase = 0.8 lowercase = 1_0 lowercase = 0.3 lowercase = 1 lowercase = 8 lowercase = 9 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = max_length lowercase = 1 lowercase = 8 lowercase = 9 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = max_length lowercase = 2 lowercase = 1 lowercase = 8 lowercase = 9 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() # pad attention mask on the left lowercase = attention_mask.at[(0, 0)].set(0) lowercase = False lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__ ,attention_mask=A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__ ,attention_mask=A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() # pad attention mask on the left lowercase = attention_mask.at[(0, 0)].set(0) lowercase = True lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__ ,attention_mask=A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__ ,attention_mask=A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() # pad attention mask on the left lowercase = attention_mask.at[(0, 0)].set(0) lowercase = 2 lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__ ,attention_mask=A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__ ,attention_mask=A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) @require_flax class lowercase ( unittest.TestCase ): def A__ ( self): lowercase = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-bert''') lowercase = FlaxAutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-bert-flax-only''') lowercase = '''Hello world''' lowercase = tokenizer(A__ ,return_tensors='''np''').input_ids # typos are quickly detected (the correct argument is `do_sample`) with self.assertRaisesRegex(A__ ,'''do_samples'''): model.generate(A__ ,do_samples=A__) # arbitrary arguments that will not be used anywhere are also not accepted with self.assertRaisesRegex(A__ ,'''foo'''): lowercase = {'''foo''': '''bar'''} model.generate(A__ ,**A__)
633
1
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_bart import BartTokenizer lowercase__ :Optional[int] = logging.get_logger(__name__) lowercase__ :int = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} # See all BART models at https://huggingface.co/models?filter=bart lowercase__ :List[str] = { "vocab_file": { "facebook/bart-base": "https://huggingface.co/facebook/bart-base/resolve/main/vocab.json", "facebook/bart-large": "https://huggingface.co/facebook/bart-large/resolve/main/vocab.json", "facebook/bart-large-mnli": "https://huggingface.co/facebook/bart-large-mnli/resolve/main/vocab.json", "facebook/bart-large-cnn": "https://huggingface.co/facebook/bart-large-cnn/resolve/main/vocab.json", "facebook/bart-large-xsum": "https://huggingface.co/facebook/bart-large-xsum/resolve/main/vocab.json", "yjernite/bart_eli5": "https://huggingface.co/yjernite/bart_eli5/resolve/main/vocab.json", }, "merges_file": { "facebook/bart-base": "https://huggingface.co/facebook/bart-base/resolve/main/merges.txt", "facebook/bart-large": "https://huggingface.co/facebook/bart-large/resolve/main/merges.txt", "facebook/bart-large-mnli": "https://huggingface.co/facebook/bart-large-mnli/resolve/main/merges.txt", "facebook/bart-large-cnn": "https://huggingface.co/facebook/bart-large-cnn/resolve/main/merges.txt", "facebook/bart-large-xsum": "https://huggingface.co/facebook/bart-large-xsum/resolve/main/merges.txt", "yjernite/bart_eli5": "https://huggingface.co/yjernite/bart_eli5/resolve/main/merges.txt", }, "tokenizer_file": { "facebook/bart-base": "https://huggingface.co/facebook/bart-base/resolve/main/tokenizer.json", "facebook/bart-large": "https://huggingface.co/facebook/bart-large/resolve/main/tokenizer.json", "facebook/bart-large-mnli": "https://huggingface.co/facebook/bart-large-mnli/resolve/main/tokenizer.json", "facebook/bart-large-cnn": "https://huggingface.co/facebook/bart-large-cnn/resolve/main/tokenizer.json", "facebook/bart-large-xsum": "https://huggingface.co/facebook/bart-large-xsum/resolve/main/tokenizer.json", "yjernite/bart_eli5": "https://huggingface.co/yjernite/bart_eli5/resolve/main/tokenizer.json", }, } lowercase__ :Optional[Any] = { "facebook/bart-base": 1024, "facebook/bart-large": 1024, "facebook/bart-large-mnli": 1024, "facebook/bart-large-cnn": 1024, "facebook/bart-large-xsum": 1024, "yjernite/bart_eli5": 1024, } class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : str =VOCAB_FILES_NAMES lowercase_ : Optional[int] =PRETRAINED_VOCAB_FILES_MAP lowercase_ : Tuple =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ : List[str] =['''input_ids''', '''attention_mask'''] lowercase_ : List[Any] =BartTokenizer def __init__( self ,A__=None ,A__=None ,A__=None ,A__="replace" ,A__="<s>" ,A__="</s>" ,A__="</s>" ,A__="<s>" ,A__="<unk>" ,A__="<pad>" ,A__="<mask>" ,A__=False ,A__=True ,**A__ ,): super().__init__( A__ ,A__ ,tokenizer_file=A__ ,errors=A__ ,bos_token=A__ ,eos_token=A__ ,sep_token=A__ ,cls_token=A__ ,unk_token=A__ ,pad_token=A__ ,mask_token=A__ ,add_prefix_space=A__ ,trim_offsets=A__ ,**A__ ,) lowercase = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__()) if pre_tok_state.get('''add_prefix_space''' ,A__) != add_prefix_space: lowercase = getattr(A__ ,pre_tok_state.pop('''type''')) lowercase = add_prefix_space lowercase = pre_tok_class(**A__) lowercase = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` lowercase = '''post_processor''' lowercase = getattr(self.backend_tokenizer ,A__ ,A__) if tokenizer_component_instance: lowercase = json.loads(tokenizer_component_instance.__getstate__()) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowercase = tuple(state['''sep''']) if "cls" in state: lowercase = tuple(state['''cls''']) lowercase = False if state.get('''add_prefix_space''' ,A__) != add_prefix_space: lowercase = add_prefix_space lowercase = True if state.get('''trim_offsets''' ,A__) != trim_offsets: lowercase = trim_offsets lowercase = True if changes_to_apply: lowercase = getattr(A__ ,state.pop('''type''')) lowercase = component_class(**A__) setattr(self.backend_tokenizer ,A__ ,A__) @property def A__ ( self): if self._mask_token is None: if self.verbose: logger.error('''Using mask_token, but it is not set yet.''') return None return str(self._mask_token) @mask_token.setter def A__ ( self ,A__): lowercase = AddedToken(A__ ,lstrip=A__ ,rstrip=A__) if isinstance(A__ ,A__) else value lowercase = value def A__ ( self ,*A__ ,**A__): lowercase = kwargs.get('''is_split_into_words''' ,A__) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' '''to use it with pretokenized inputs.''') return super()._batch_encode_plus(*A__ ,**A__) def A__ ( self ,*A__ ,**A__): lowercase = kwargs.get('''is_split_into_words''' ,A__) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' '''to use it with pretokenized inputs.''') return super()._encode_plus(*A__ ,**A__) def A__ ( self ,A__ ,A__ = None): lowercase = self._tokenizer.model.save(A__ ,name=A__) return tuple(A__) def A__ ( self ,A__ ,A__=None): lowercase = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def A__ ( self ,A__ ,A__ = None): lowercase = [self.sep_token_id] lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0]
633
import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : List[str] =(UniPCMultistepScheduler,) lowercase_ : Tuple =(('''num_inference_steps''', 25),) def A__ ( self ,**A__): lowercase = { '''num_train_timesteps''': 1_0_0_0, '''beta_start''': 0.0001, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', '''solver_order''': 2, '''solver_type''': '''bh2''', } config.update(**A__) return config def A__ ( self ,A__=0 ,**A__): lowercase = dict(self.forward_default_kwargs) lowercase = kwargs.pop('''num_inference_steps''' ,A__) lowercase = self.dummy_sample lowercase = 0.1 * sample lowercase = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) scheduler.set_timesteps(A__) # copy over dummy past residuals lowercase = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A__) lowercase = scheduler_class.from_pretrained(A__) new_scheduler.set_timesteps(A__) # copy over dummy past residuals lowercase = dummy_past_residuals[: new_scheduler.config.solver_order] lowercase , lowercase = sample, sample for t in range(A__ ,time_step + scheduler.config.solver_order + 1): lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = new_scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1E-5, "Scheduler outputs are not identical" def A__ ( self ,A__=0 ,**A__): lowercase = dict(self.forward_default_kwargs) lowercase = kwargs.pop('''num_inference_steps''' ,A__) lowercase = self.dummy_sample lowercase = 0.1 * sample lowercase = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config() lowercase = scheduler_class(**A__) scheduler.set_timesteps(A__) # copy over dummy past residuals (must be after setting timesteps) lowercase = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A__) lowercase = scheduler_class.from_pretrained(A__) # copy over dummy past residuals new_scheduler.set_timesteps(A__) # copy over dummy past residual (must be after setting timesteps) lowercase = dummy_past_residuals[: new_scheduler.config.solver_order] lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = new_scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1E-5, "Scheduler outputs are not identical" def A__ ( self ,A__=None ,**A__): if scheduler is None: lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) lowercase = 1_0 lowercase = self.dummy_model() lowercase = self.dummy_sample_deter scheduler.set_timesteps(A__) for i, t in enumerate(scheduler.timesteps): lowercase = model(A__ ,A__) lowercase = scheduler.step(A__ ,A__ ,A__).prev_sample return sample def A__ ( self): lowercase = dict(self.forward_default_kwargs) lowercase = kwargs.pop('''num_inference_steps''' ,A__) for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config() lowercase = scheduler_class(**A__) lowercase = self.dummy_sample lowercase = 0.1 * sample if num_inference_steps is not None and hasattr(A__ ,'''set_timesteps'''): scheduler.set_timesteps(A__) elif num_inference_steps is not None and not hasattr(A__ ,'''set_timesteps'''): lowercase = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) lowercase = [residual + 0.2, residual + 0.15, residual + 0.10] lowercase = dummy_past_residuals[: scheduler.config.solver_order] lowercase = scheduler.timesteps[5] lowercase = scheduler.timesteps[6] lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample self.assertEqual(output_a.shape ,sample.shape) self.assertEqual(output_a.shape ,output_a.shape) def A__ ( self): # make sure that iterating over schedulers with same config names gives same results # for defaults lowercase = UniPCMultistepScheduler(**self.get_scheduler_config()) lowercase = self.full_loop(scheduler=A__) lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.2464) < 1E-3 lowercase = DPMSolverSinglestepScheduler.from_config(scheduler.config) lowercase = DEISMultistepScheduler.from_config(scheduler.config) lowercase = DPMSolverMultistepScheduler.from_config(scheduler.config) lowercase = UniPCMultistepScheduler.from_config(scheduler.config) lowercase = self.full_loop(scheduler=A__) lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.2464) < 1E-3 def A__ ( self): for timesteps in [2_5, 5_0, 1_0_0, 9_9_9, 1_0_0_0]: self.check_over_configs(num_train_timesteps=A__) def A__ ( self): self.check_over_configs(thresholding=A__) for order in [1, 2, 3]: for solver_type in ["bh1", "bh2"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=A__ ,prediction_type=A__ ,sample_max_value=A__ ,solver_order=A__ ,solver_type=A__ ,) def A__ ( self): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=A__) def A__ ( self): for solver_type in ["bh1", "bh2"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=A__ ,solver_type=A__ ,prediction_type=A__ ,) lowercase = self.full_loop( solver_order=A__ ,solver_type=A__ ,prediction_type=A__ ,) assert not torch.isnan(A__).any(), "Samples have nan numbers" def A__ ( self): self.check_over_configs(lower_order_final=A__) self.check_over_configs(lower_order_final=A__) def A__ ( self): for num_inference_steps in [1, 2, 3, 5, 1_0, 5_0, 1_0_0, 9_9_9, 1_0_0_0]: self.check_over_forward(num_inference_steps=A__ ,time_step=0) def A__ ( self): lowercase = self.full_loop() lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.2464) < 1E-3 def A__ ( self): lowercase = self.full_loop(prediction_type='''v_prediction''') lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.1014) < 1E-3 def A__ ( self): lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config(thresholding=A__ ,dynamic_thresholding_ratio=0) lowercase = scheduler_class(**A__) lowercase = 1_0 lowercase = self.dummy_model() lowercase = self.dummy_sample_deter.half() scheduler.set_timesteps(A__) for i, t in enumerate(scheduler.timesteps): lowercase = model(A__ ,A__) lowercase = scheduler.step(A__ ,A__ ,A__).prev_sample assert sample.dtype == torch.floataa def A__ ( self ,**A__): for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) scheduler.set_timesteps(scheduler.config.num_train_timesteps) assert len(scheduler.timesteps.unique()) == scheduler.num_inference_steps
633
1
def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' return round(float(moles / volume ) * nfactor ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' return round(float((moles * 0.08_21 * temperature) / (volume) ) ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' return round(float((moles * 0.08_21 * temperature) / (pressure) ) ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' return round(float((pressure * volume) / (0.08_21 * moles) ) ) if __name__ == "__main__": import doctest doctest.testmod()
633
import argparse from collections import defaultdict import yaml lowercase__ :Optional[int] = "docs/source/en/_toctree.yml" def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = defaultdict(lowerCAmelCase__ ) for doc in model_doc: counts[doc["local"]] += 1 lowercase = [key for key, value in counts.items() if value > 1] lowercase = [] for duplicate_key in duplicates: lowercase = list({doc['''title'''] for doc in model_doc if doc['''local'''] == duplicate_key} ) if len(lowerCAmelCase__ ) > 1: raise ValueError( f'{duplicate_key} is present several times in the documentation table of content at ' '''`docs/source/en/_toctree.yml` with different *Title* values. Choose one of those and remove the ''' '''others.''' ) # Only add this once new_doc.append({'''local''': duplicate_key, '''title''': titles[0]} ) # Add none duplicate-keys new_doc.extend([doc for doc in model_doc if counts[doc['''local''']] == 1] ) # Sort return sorted(lowerCAmelCase__ , key=lambda lowerCAmelCase__ : s["title"].lower() ) def UpperCamelCase ( lowerCAmelCase__=False ): '''simple docstring''' with open(lowerCAmelCase__ , encoding='''utf-8''' ) as f: lowercase = yaml.safe_load(f.read() ) # Get to the API doc lowercase = 0 while content[api_idx]["title"] != "API": api_idx += 1 lowercase = content[api_idx]['''sections'''] # Then to the model doc lowercase = 0 while api_doc[model_idx]["title"] != "Models": model_idx += 1 lowercase = api_doc[model_idx]['''sections'''] lowercase = [(idx, section) for idx, section in enumerate(lowerCAmelCase__ ) if '''sections''' in section] lowercase = False for idx, modality_doc in modalities_docs: lowercase = modality_doc['''sections'''] lowercase = clean_model_doc_toc(lowerCAmelCase__ ) if old_modality_doc != new_modality_doc: lowercase = True if overwrite: lowercase = new_modality_doc if diff: if overwrite: lowercase = model_doc lowercase = api_doc with open(lowerCAmelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(yaml.dump(lowerCAmelCase__ , allow_unicode=lowerCAmelCase__ ) ) else: raise ValueError( '''The model doc part of the table of content is not properly sorted, run `make style` to fix this.''' ) if __name__ == "__main__": lowercase__ :Any = argparse.ArgumentParser() parser.add_argument("--fix_and_overwrite", action="store_true", help="Whether to fix inconsistencies.") lowercase__ :int = parser.parse_args() check_model_doc(args.fix_and_overwrite)
633
1
import logging from transformers import PretrainedConfig lowercase__ :int = logging.getLogger(__name__) lowercase__ :Dict = { "bertabs-finetuned-cnndm": "https://huggingface.co/remi/bertabs-finetuned-cnndm-extractive-abstractive-summarization/resolve/main/config.json", } class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Optional[int] ='''bertabs''' def __init__( self ,A__=3_0_5_2_2 ,A__=5_1_2 ,A__=6 ,A__=5_1_2 ,A__=8 ,A__=5_1_2 ,A__=0.2 ,A__=6 ,A__=7_6_8 ,A__=8 ,A__=2_0_4_8 ,A__=0.2 ,**A__ ,): super().__init__(**A__) lowercase = vocab_size lowercase = max_pos lowercase = enc_layers lowercase = enc_hidden_size lowercase = enc_heads lowercase = enc_ff_size lowercase = enc_dropout lowercase = dec_layers lowercase = dec_hidden_size lowercase = dec_heads lowercase = dec_ff_size lowercase = dec_dropout
633
from __future__ import annotations from decimal import Decimal from math import * # noqa: F403 from sympy import diff def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 10**-10 ): '''simple docstring''' lowercase = a while True: lowercase = Decimal(lowerCAmelCase__ ) - ( Decimal(eval(lowerCAmelCase__ ) ) / Decimal(eval(str(diff(lowerCAmelCase__ ) ) ) ) # noqa: S307 ) # This number dictates the accuracy of the answer if abs(eval(lowerCAmelCase__ ) ) < precision: # noqa: S307 return float(lowerCAmelCase__ ) # Let's Execute if __name__ == "__main__": # Find root of trigonometric function # Find value of pi print(F'The root of sin(x) = 0 is {newton_raphson("sin(x)", 2)}') # Find root of polynomial print(F'The root of x**2 - 5*x + 2 = 0 is {newton_raphson("x**2 - 5*x + 2", 0.4)}') # Find Square Root of 5 print(F'The root of log(x) - 1 = 0 is {newton_raphson("log(x) - 1", 2)}') # Exponential Roots print(F'The root of exp(x) - 1 = 0 is {newton_raphson("exp(x) - 1", 0)}')
633
1
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowercase__ :List[str] = logging.get_logger(__name__) class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Dict =['''pixel_values'''] def __init__( self ,A__ = True ,A__ = None ,A__ = PIL.Image.BICUBIC ,A__ = True ,A__ = None ,A__ = 1 / 2_5_5 ,A__ = True ,A__ = True ,A__ = None ,A__ = None ,**A__ ,): super().__init__(**A__) lowercase = size if size is not None else {'''height''': 2_5_6, '''width''': 2_5_6} lowercase = get_size_dict(A__) lowercase = crop_size if crop_size is not None else {'''height''': 2_2_4, '''width''': 2_2_4} lowercase = get_size_dict(A__ ,param_name='''crop_size''') lowercase = do_resize lowercase = size lowercase = resample lowercase = do_center_crop lowercase = crop_size lowercase = do_rescale lowercase = rescale_factor lowercase = do_normalize lowercase = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN lowercase = image_std if image_std is not None else IMAGENET_STANDARD_STD def A__ ( self ,A__ ,A__ ,A__ = PIL.Image.BICUBIC ,A__ = None ,**A__ ,): lowercase = get_size_dict(A__) if "height" not in size or "width" not in size: raise ValueError(f'The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}') return resize( A__ ,size=(size['''height'''], size['''width''']) ,resample=A__ ,data_format=A__ ,**A__) def A__ ( self ,A__ ,A__ ,A__ = None ,**A__ ,): lowercase = get_size_dict(A__) if "height" not in size or "width" not in size: raise ValueError(f'The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}') return center_crop(A__ ,size=(size['''height'''], size['''width''']) ,data_format=A__ ,**A__) def A__ ( self ,A__ ,A__ ,A__ = None ,**A__ ,): return rescale(A__ ,scale=A__ ,data_format=A__ ,**A__) def A__ ( self ,A__ ,A__ ,A__ ,A__ = None ,**A__ ,): return normalize(A__ ,mean=A__ ,std=A__ ,data_format=A__ ,**A__) def A__ ( self ,A__ ,A__ = None ,A__ = None ,A__=None ,A__ = None ,A__ = None ,A__ = None ,A__ = None ,A__ = None ,A__ = None ,A__ = None ,A__ = None ,A__ = ChannelDimension.FIRST ,**A__ ,): lowercase = do_resize if do_resize is not None else self.do_resize lowercase = resample if resample is not None else self.resample lowercase = do_center_crop if do_center_crop is not None else self.do_center_crop lowercase = do_rescale if do_rescale is not None else self.do_rescale lowercase = rescale_factor if rescale_factor is not None else self.rescale_factor lowercase = do_normalize if do_normalize is not None else self.do_normalize lowercase = image_mean if image_mean is not None else self.image_mean lowercase = image_std if image_std is not None else self.image_std lowercase = size if size is not None else self.size lowercase = get_size_dict(A__) lowercase = crop_size if crop_size is not None else self.crop_size lowercase = get_size_dict(A__ ,param_name='''crop_size''') lowercase = make_list_of_images(A__) if not valid_images(A__): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''') if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''') if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''') if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''') if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''') # All transformations expect numpy arrays. lowercase = [to_numpy_array(A__) for image in images] if do_resize: lowercase = [self.resize(image=A__ ,size=A__ ,resample=A__) for image in images] if do_center_crop: lowercase = [self.center_crop(image=A__ ,size=A__) for image in images] if do_rescale: lowercase = [self.rescale(image=A__ ,scale=A__) for image in images] if do_normalize: lowercase = [self.normalize(image=A__ ,mean=A__ ,std=A__) for image in images] lowercase = [to_channel_dimension_format(A__ ,A__) for image in images] lowercase = {'''pixel_values''': images} return BatchFeature(data=A__ ,tensor_type=A__)
633
from pathlib import Path import numpy as np from PIL import Image def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase , lowercase = rgb[:, :, 0], rgb[:, :, 1], rgb[:, :, 2] return 0.29_89 * r + 0.58_70 * g + 0.11_40 * b def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return (gray > 127) & (gray <= 255) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = np.zeros_like(lowerCAmelCase__ ) lowercase = np.zeros( (image.shape[0] + kernel.shape[0] - 1, image.shape[1] + kernel.shape[1] - 1) ) # Copy image to padded image lowercase = image # Iterate over image & apply kernel for x in range(image.shape[1] ): for y in range(image.shape[0] ): lowercase = ( kernel * image_padded[y : y + kernel.shape[0], x : x + kernel.shape[1]] ).sum() lowercase = int(summation > 0 ) return output if __name__ == "__main__": # read original image lowercase__ :str = Path(__file__).resolve().parent / "image_data" / "lena.jpg" lowercase__ :List[str] = np.array(Image.open(lena_path)) # kernel to be applied lowercase__ :Union[str, Any] = np.array([[0, 1, 0], [1, 1, 1], [0, 1, 0]]) lowercase__ :Optional[int] = dilation(gray_to_binary(rgb_to_gray(lena)), structuring_element) # Save the output image lowercase__ :str = Image.fromarray(output).convert("RGB") pil_img.save("result_dilation.png")
633
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowercase__ :Dict = { "configuration_mask2former": [ "MASK2FORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "Mask2FormerConfig", ], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :Union[str, Any] = ["Mask2FormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :Any = [ "MASK2FORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "Mask2FormerForUniversalSegmentation", "Mask2FormerModel", "Mask2FormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_maskaformer import MASK2FORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, MaskaFormerConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_maskaformer import MaskaFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_maskaformer import ( MASK2FORMER_PRETRAINED_MODEL_ARCHIVE_LIST, MaskaFormerForUniversalSegmentation, MaskaFormerModel, MaskaFormerPreTrainedModel, ) else: import sys lowercase__ :List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure)
633
import argparse import torch from torch import nn from transformers import MaMaaaConfig, MaMaaaForConditionalGeneration def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = [ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''decoder.output_projection.weight''', '''_float_tensor''', '''encoder.embed_positions._float_tensor''', '''decoder.embed_positions._float_tensor''', ] for k in ignore_keys: state_dict.pop(lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase = emb.weight.shape lowercase = nn.Linear(lowerCAmelCase__ , lowerCAmelCase__ , bias=lowerCAmelCase__ ) lowercase = emb.weight.data return lin_layer def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = torch.load(lowerCAmelCase__ , map_location='''cpu''' ) lowercase = mam_aaa['''args'''] or mam_aaa['''cfg''']['''model'''] lowercase = mam_aaa['''model'''] remove_ignore_keys_(lowerCAmelCase__ ) lowercase = state_dict['''encoder.embed_tokens.weight'''].shape[0] lowercase = MaMaaaConfig( vocab_size=lowerCAmelCase__ , max_position_embeddings=1024 , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , encoder_layerdrop=args.encoder_layerdrop , decoder_layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='''relu''' , ) lowercase = state_dict['''decoder.embed_tokens.weight'''] lowercase = MaMaaaForConditionalGeneration(lowerCAmelCase__ ) model.model.load_state_dict(lowerCAmelCase__ , strict=lowerCAmelCase__ ) lowercase = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": lowercase__ :Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument("fairseq_path", type=str, help="path to a model.pt on local filesystem.") parser.add_argument("pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") lowercase__ :Tuple = parser.parse_args() lowercase__ :int = convert_fairseq_mamaaa_checkpoint_from_disk(args.fairseq_pathß) model.save_pretrained(args.pytorch_dump_folder_path)
633
1
import argparse import hashlib # hashlib is only used inside the Test class import struct class lowercase : def __init__( self ,A__): lowercase = data lowercase = [0X67_45_23_01, 0Xef_cd_ab_89, 0X98_ba_dc_fe, 0X10_32_54_76, 0Xc3_d2_e1_f0] @staticmethod def A__ ( A__ ,A__): return ((n << b) | (n >> (3_2 - b))) & 0Xff_ff_ff_ff def A__ ( self): lowercase = B'''\x80''' + B'''\x00''' * (6_3 - (len(self.data) + 8) % 6_4) lowercase = self.data + padding + struct.pack('''>Q''' ,8 * len(self.data)) return padded_data def A__ ( self): return [ self.padded_data[i : i + 6_4] for i in range(0 ,len(self.padded_data) ,6_4) ] def A__ ( self ,A__): lowercase = list(struct.unpack('''>16L''' ,A__)) + [0] * 6_4 for i in range(1_6 ,8_0): lowercase = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 1_4] ^ w[i - 1_6]) ,1) return w def A__ ( self): lowercase = self.padding() lowercase = self.split_blocks() for block in self.blocks: lowercase = self.expand_block(A__) lowercase , lowercase , lowercase , lowercase , lowercase = self.h for i in range(0 ,8_0): if 0 <= i < 2_0: lowercase = (b & c) | ((~b) & d) lowercase = 0X5a_82_79_99 elif 2_0 <= i < 4_0: lowercase = b ^ c ^ d lowercase = 0X6e_d9_eb_a1 elif 4_0 <= i < 6_0: lowercase = (b & c) | (b & d) | (c & d) lowercase = 0X8f_1b_bc_dc elif 6_0 <= i < 8_0: lowercase = b ^ c ^ d lowercase = 0Xca_62_c1_d6 lowercase , lowercase , lowercase , lowercase , lowercase = ( self.rotate(A__ ,5) + f + e + k + expanded_block[i] & 0Xff_ff_ff_ff, a, self.rotate(A__ ,3_0), c, d, ) lowercase = ( self.h[0] + a & 0Xff_ff_ff_ff, self.h[1] + b & 0Xff_ff_ff_ff, self.h[2] + c & 0Xff_ff_ff_ff, self.h[3] + d & 0Xff_ff_ff_ff, self.h[4] + e & 0Xff_ff_ff_ff, ) return ("{:08x}" * 5).format(*self.h) def UpperCamelCase ( ): '''simple docstring''' lowercase = B'''Test String''' assert SHAaHash(lowerCAmelCase__ ).final_hash() == hashlib.shaa(lowerCAmelCase__ ).hexdigest() # noqa: S324 def UpperCamelCase ( ): '''simple docstring''' lowercase = argparse.ArgumentParser(description='''Process some strings or files''' ) parser.add_argument( '''--string''' , dest='''input_string''' , default='''Hello World!! Welcome to Cryptography''' , help='''Hash the string''' , ) parser.add_argument('''--file''' , dest='''input_file''' , help='''Hash contents of a file''' ) lowercase = parser.parse_args() lowercase = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file , '''rb''' ) as f: lowercase = f.read() else: lowercase = bytes(lowerCAmelCase__ , '''utf-8''' ) print(SHAaHash(lowerCAmelCase__ ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
633
from __future__ import annotations from random import random class lowercase : def __init__( self ,A__ = None): lowercase = value lowercase = random() lowercase = None lowercase = None def __repr__( self): from pprint import pformat if self.left is None and self.right is None: return f'\'{self.value}: {self.prior:.5}\'' else: return pformat( {f'{self.value}: {self.prior:.5}': (self.left, self.right)} ,indent=1) def __str__( self): lowercase = str(self.value) + ''' ''' lowercase = str(self.left or '''''') lowercase = str(self.right or '''''') return value + left + right def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' if root is None: # None tree is split into 2 Nones return None, None elif root.value is None: return None, None else: if value < root.value: lowercase , lowercase = split(root.left , lowerCAmelCase__ ) return left, root else: lowercase , lowercase = split(root.right , lowerCAmelCase__ ) return root, right def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' if (not left) or (not right): # If one node is None, return the other return left or right elif left.prior < right.prior: lowercase = merge(left.right , lowerCAmelCase__ ) return left else: lowercase = merge(lowerCAmelCase__ , right.left ) return right def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = Node(lowerCAmelCase__ ) lowercase , lowercase = split(lowerCAmelCase__ , lowerCAmelCase__ ) return merge(merge(lowerCAmelCase__ , lowerCAmelCase__ ) , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase = split(lowerCAmelCase__ , value - 1 ) lowercase , lowercase = split(lowerCAmelCase__ , lowerCAmelCase__ ) return merge(lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if not root: # None return else: inorder(root.left ) print(root.value , end=''',''' ) inorder(root.right ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' for arg in args.split(): if arg[0] == "+": lowercase = insert(lowerCAmelCase__ , int(arg[1:] ) ) elif arg[0] == "-": lowercase = erase(lowerCAmelCase__ , int(arg[1:] ) ) else: print('''Unknown command''' ) return root def UpperCamelCase ( ): '''simple docstring''' lowercase = None print( '''enter numbers to create a tree, + value to add value into treap, ''' '''- value to erase all nodes with value. \'q\' to quit. ''' ) lowercase = input() while args != "q": lowercase = interact_treap(lowerCAmelCase__ , lowerCAmelCase__ ) print(lowerCAmelCase__ ) lowercase = input() print('''good by!''' ) if __name__ == "__main__": import doctest doctest.testmod() main()
633
1
import argparse import os import torch from transformers.utils import WEIGHTS_NAME lowercase__ :List[Any] = ["small", "medium", "large"] lowercase__ :List[Any] = "lm_head.decoder.weight" lowercase__ :str = "lm_head.weight" def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = torch.load(lowerCAmelCase__ ) lowercase = d.pop(lowerCAmelCase__ ) os.makedirs(lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) torch.save(lowerCAmelCase__ , os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) ) if __name__ == "__main__": lowercase__ :Dict = argparse.ArgumentParser() parser.add_argument("--dialogpt_path", default=".", type=str) lowercase__ :Optional[int] = parser.parse_args() for MODEL in DIALOGPT_MODELS: lowercase__ :List[str] = os.path.join(args.dialogpt_path, F'{MODEL}_ft.pkl') lowercase__ :Union[str, Any] = F'./DialoGPT-{MODEL}' convert_dialogpt_checkpoint( checkpoint_path, pytorch_dump_folder_path, )
633
def UpperCamelCase ( lowerCAmelCase__ = 1000 ): '''simple docstring''' lowercase = -1 lowercase = 0 for a in range(1 , n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N eliminating c lowercase = (n * n - 2 * a * n) // (2 * n - 2 * a) lowercase = n - a - b if c * c == (a * a + b * b): lowercase = a * b * c if candidate >= product: lowercase = candidate return product if __name__ == "__main__": print(F'{solution() = }')
633
1
import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Optional[Any] =(PNDMScheduler,) lowercase_ : List[Any] =(('''num_inference_steps''', 50),) def A__ ( self ,**A__): lowercase = { '''num_train_timesteps''': 1_0_0_0, '''beta_start''': 0.0001, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', } config.update(**A__) return config def A__ ( self ,A__=0 ,**A__): lowercase = dict(self.forward_default_kwargs) lowercase = kwargs.pop('''num_inference_steps''' ,A__) lowercase = self.dummy_sample lowercase = 0.1 * sample lowercase = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) scheduler.set_timesteps(A__) # copy over dummy past residuals lowercase = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A__) lowercase = scheduler_class.from_pretrained(A__) new_scheduler.set_timesteps(A__) # copy over dummy past residuals lowercase = dummy_past_residuals[:] lowercase = scheduler.step_prk(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = new_scheduler.step_prk(A__ ,A__ ,A__ ,**A__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1E-5, "Scheduler outputs are not identical" lowercase = scheduler.step_plms(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = new_scheduler.step_plms(A__ ,A__ ,A__ ,**A__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1E-5, "Scheduler outputs are not identical" def A__ ( self): pass def A__ ( self ,A__=0 ,**A__): lowercase = dict(self.forward_default_kwargs) lowercase = kwargs.pop('''num_inference_steps''' ,A__) lowercase = self.dummy_sample lowercase = 0.1 * sample lowercase = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config() lowercase = scheduler_class(**A__) scheduler.set_timesteps(A__) # copy over dummy past residuals (must be after setting timesteps) lowercase = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A__) lowercase = scheduler_class.from_pretrained(A__) # copy over dummy past residuals new_scheduler.set_timesteps(A__) # copy over dummy past residual (must be after setting timesteps) lowercase = dummy_past_residuals[:] lowercase = scheduler.step_prk(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = new_scheduler.step_prk(A__ ,A__ ,A__ ,**A__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1E-5, "Scheduler outputs are not identical" lowercase = scheduler.step_plms(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = new_scheduler.step_plms(A__ ,A__ ,A__ ,**A__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1E-5, "Scheduler outputs are not identical" def A__ ( self ,**A__): lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) lowercase = 1_0 lowercase = self.dummy_model() lowercase = self.dummy_sample_deter scheduler.set_timesteps(A__) for i, t in enumerate(scheduler.prk_timesteps): lowercase = model(A__ ,A__) lowercase = scheduler.step_prk(A__ ,A__ ,A__).prev_sample for i, t in enumerate(scheduler.plms_timesteps): lowercase = model(A__ ,A__) lowercase = scheduler.step_plms(A__ ,A__ ,A__).prev_sample return sample def A__ ( self): lowercase = dict(self.forward_default_kwargs) lowercase = kwargs.pop('''num_inference_steps''' ,A__) for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config() lowercase = scheduler_class(**A__) lowercase = self.dummy_sample lowercase = 0.1 * sample if num_inference_steps is not None and hasattr(A__ ,'''set_timesteps'''): scheduler.set_timesteps(A__) elif num_inference_steps is not None and not hasattr(A__ ,'''set_timesteps'''): lowercase = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) lowercase = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] lowercase = dummy_past_residuals[:] lowercase = scheduler.step_prk(A__ ,0 ,A__ ,**A__).prev_sample lowercase = scheduler.step_prk(A__ ,1 ,A__ ,**A__).prev_sample self.assertEqual(output_a.shape ,sample.shape) self.assertEqual(output_a.shape ,output_a.shape) lowercase = scheduler.step_plms(A__ ,0 ,A__ ,**A__).prev_sample lowercase = scheduler.step_plms(A__ ,1 ,A__ ,**A__).prev_sample self.assertEqual(output_a.shape ,sample.shape) self.assertEqual(output_a.shape ,output_a.shape) def A__ ( self): for timesteps in [1_0_0, 1_0_0_0]: self.check_over_configs(num_train_timesteps=A__) def A__ ( self): for steps_offset in [0, 1]: self.check_over_configs(steps_offset=A__) lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config(steps_offset=1) lowercase = scheduler_class(**A__) scheduler.set_timesteps(1_0) assert torch.equal( scheduler.timesteps ,torch.LongTensor( [9_0_1, 8_5_1, 8_5_1, 8_0_1, 8_0_1, 7_5_1, 7_5_1, 7_0_1, 7_0_1, 6_5_1, 6_5_1, 6_0_1, 6_0_1, 5_0_1, 4_0_1, 3_0_1, 2_0_1, 1_0_1, 1]) ,) def A__ ( self): for beta_start, beta_end in zip([0.0001, 0.001] ,[0.002, 0.02]): self.check_over_configs(beta_start=A__ ,beta_end=A__) def A__ ( self): for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=A__) def A__ ( self): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=A__) def A__ ( self): for t in [1, 5, 1_0]: self.check_over_forward(time_step=A__) def A__ ( self): for t, num_inference_steps in zip([1, 5, 1_0] ,[1_0, 5_0, 1_0_0]): self.check_over_forward(num_inference_steps=A__) def A__ ( self): # earlier version of set_timesteps() caused an error indexing alpha's with inference steps as power of 3 lowercase = 2_7 for scheduler_class in self.scheduler_classes: lowercase = self.dummy_sample lowercase = 0.1 * sample lowercase = self.get_scheduler_config() lowercase = scheduler_class(**A__) scheduler.set_timesteps(A__) # before power of 3 fix, would error on first step, so we only need to do two for i, t in enumerate(scheduler.prk_timesteps[:2]): lowercase = scheduler.step_prk(A__ ,A__ ,A__).prev_sample def A__ ( self): with self.assertRaises(A__): lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config() lowercase = scheduler_class(**A__) scheduler.step_plms(self.dummy_sample ,1 ,self.dummy_sample).prev_sample def A__ ( self): lowercase = self.full_loop() lowercase = torch.sum(torch.abs(A__)) lowercase = torch.mean(torch.abs(A__)) assert abs(result_sum.item() - 198.1318) < 1E-2 assert abs(result_mean.item() - 0.2580) < 1E-3 def A__ ( self): lowercase = self.full_loop(prediction_type='''v_prediction''') lowercase = torch.sum(torch.abs(A__)) lowercase = torch.mean(torch.abs(A__)) assert abs(result_sum.item() - 67.3986) < 1E-2 assert abs(result_mean.item() - 0.0878) < 1E-3 def A__ ( self): # We specify different beta, so that the first alpha is 0.99 lowercase = self.full_loop(set_alpha_to_one=A__ ,beta_start=0.01) lowercase = torch.sum(torch.abs(A__)) lowercase = torch.mean(torch.abs(A__)) assert abs(result_sum.item() - 230.0399) < 1E-2 assert abs(result_mean.item() - 0.2995) < 1E-3 def A__ ( self): # We specify different beta, so that the first alpha is 0.99 lowercase = self.full_loop(set_alpha_to_one=A__ ,beta_start=0.01) lowercase = torch.sum(torch.abs(A__)) lowercase = torch.mean(torch.abs(A__)) assert abs(result_sum.item() - 186.9482) < 1E-2 assert abs(result_mean.item() - 0.2434) < 1E-3
633
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ :Tuple = { "configuration_instructblip": [ "INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "InstructBlipConfig", "InstructBlipQFormerConfig", "InstructBlipVisionConfig", ], "processing_instructblip": ["InstructBlipProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :List[str] = [ "INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "InstructBlipQFormerModel", "InstructBlipPreTrainedModel", "InstructBlipForConditionalGeneration", "InstructBlipVisionModel", ] if TYPE_CHECKING: from .configuration_instructblip import ( INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, InstructBlipConfig, InstructBlipQFormerConfig, InstructBlipVisionConfig, ) from .processing_instructblip import InstructBlipProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_instructblip import ( INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST, InstructBlipForConditionalGeneration, InstructBlipPreTrainedModel, InstructBlipQFormerModel, InstructBlipVisionModel, ) else: import sys lowercase__ :List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
633
1
from __future__ import annotations from collections.abc import Callable from typing import Generic, TypeVar lowercase__ :List[Any] = TypeVar("T") lowercase__ :Optional[Any] = TypeVar("U") class lowercase ( Generic[T, U] ): def __init__( self ,A__ ,A__): lowercase = key lowercase = val lowercase = None lowercase = None def __repr__( self): return ( f'Node: key: {self.key}, val: {self.val}, ' f'has next: {bool(self.next)}, has prev: {bool(self.prev)}' ) class lowercase ( Generic[T, U] ): def __init__( self): lowercase = DoubleLinkedListNode(A__ ,A__) lowercase = DoubleLinkedListNode(A__ ,A__) lowercase , lowercase = self.rear, self.head def __repr__( self): lowercase = ['''DoubleLinkedList'''] lowercase = self.head while node.next is not None: rep.append(str(A__)) lowercase = node.next rep.append(str(self.rear)) return ",\n ".join(A__) def A__ ( self ,A__): lowercase = self.rear.prev # All nodes other than self.head are guaranteed to have non-None previous assert previous is not None lowercase = node lowercase = previous lowercase = node lowercase = self.rear def A__ ( self ,A__): if node.prev is None or node.next is None: return None lowercase = node.next lowercase = node.prev lowercase = None lowercase = None return node class lowercase ( Generic[T, U] ): lowercase_ : dict[Callable[[T], U], LRUCache[T, U]] ={} def __init__( self ,A__): lowercase = DoubleLinkedList() lowercase = capacity lowercase = 0 lowercase = 0 lowercase = 0 lowercase = {} def __repr__( self): return ( f'CacheInfo(hits={self.hits}, misses={self.miss}, ' f'capacity={self.capacity}, current size={self.num_keys})' ) def __contains__( self ,A__): return key in self.cache def A__ ( self ,A__): # Note: pythonic interface would throw KeyError rather than return None if key in self.cache: self.hits += 1 lowercase = self.cache[key] lowercase = self.list.remove(self.cache[key]) assert node == value_node # node is guaranteed not None because it is in self.cache assert node is not None self.list.add(A__) return node.val self.miss += 1 return None def A__ ( self ,A__ ,A__): if key not in self.cache: if self.num_keys >= self.capacity: # delete first node (oldest) when over capacity lowercase = self.list.head.next # guaranteed to have a non-None first node when num_keys > 0 # explain to type checker via assertions assert first_node is not None assert first_node.key is not None assert ( self.list.remove(A__) is not None ) # node guaranteed to be in list assert node.key is not None del self.cache[first_node.key] self.num_keys -= 1 lowercase = DoubleLinkedListNode(A__ ,A__) self.list.add(self.cache[key]) self.num_keys += 1 else: # bump node to the end of the list, update value lowercase = self.list.remove(self.cache[key]) assert node is not None # node guaranteed to be in list lowercase = value self.list.add(A__) @classmethod def A__ ( cls ,A__ = 1_2_8): def cache_decorator_inner(A__) -> Callable[..., U]: def cache_decorator_wrapper(*A__) -> U: if func not in cls.decorator_function_to_instance_map: lowercase = LRUCache(A__) lowercase = cls.decorator_function_to_instance_map[func].get(args[0]) if result is None: lowercase = func(*A__) cls.decorator_function_to_instance_map[func].put(args[0] ,A__) return result def cache_info() -> LRUCache[T, U]: return cls.decorator_function_to_instance_map[func] setattr(A__ ,'''cache_info''' ,A__) # noqa: B010 return cache_decorator_wrapper return cache_decorator_inner if __name__ == "__main__": import doctest doctest.testmod()
633
from numpy import exp, pi, sqrt def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ = 0.0 , lowerCAmelCase__ = 1.0 ): '''simple docstring''' return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
633
1
from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ :List[str] = logging.get_logger(__name__) lowercase__ :Dict = { "EleutherAI/gpt-neox-20b": "https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/config.json", # See all GPTNeoX models at https://huggingface.co/models?filter=gpt_neox } class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Tuple ='''gpt_neox''' def __init__( self ,A__=5_0_4_3_2 ,A__=6_1_4_4 ,A__=4_4 ,A__=6_4 ,A__=2_4_5_7_6 ,A__="gelu" ,A__=0.25 ,A__=1_0_0_0_0 ,A__=0.0 ,A__=0.0 ,A__=0.1 ,A__=2_0_4_8 ,A__=0.02 ,A__=1E-5 ,A__=True ,A__=0 ,A__=2 ,A__=False ,A__=True ,A__=None ,**A__ ,): super().__init__(bos_token_id=A__ ,eos_token_id=A__ ,**A__) lowercase = vocab_size lowercase = max_position_embeddings lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_attention_heads lowercase = intermediate_size lowercase = hidden_act lowercase = rotary_pct lowercase = rotary_emb_base lowercase = attention_dropout lowercase = hidden_dropout lowercase = classifier_dropout lowercase = initializer_range lowercase = layer_norm_eps lowercase = use_cache lowercase = tie_word_embeddings lowercase = use_parallel_residual lowercase = rope_scaling self._rope_scaling_validation() if self.hidden_size % self.num_attention_heads != 0: raise ValueError( '''The hidden size is not divisble by the number of attention heads! Make sure to update them!''') def A__ ( self): if self.rope_scaling is None: return if not isinstance(self.rope_scaling ,A__) or len(self.rope_scaling) != 2: raise ValueError( '''`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, ''' f'got {self.rope_scaling}') lowercase = self.rope_scaling.get('''type''' ,A__) lowercase = self.rope_scaling.get('''factor''' ,A__) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f'`rope_scaling`\'s name field must be one of [\'linear\', \'dynamic\'], got {rope_scaling_type}') if rope_scaling_factor is None or not isinstance(A__ ,A__) or rope_scaling_factor <= 1.0: raise ValueError(f'`rope_scaling`\'s factor field must be an float > 1, got {rope_scaling_factor}')
633
import asyncio import os import shutil import subprocess import sys import tempfile import unittest from distutils.util import strtobool from functools import partial from pathlib import Path from typing import List, Union from unittest import mock import torch from ..state import AcceleratorState, PartialState from ..utils import ( gather, is_bnb_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_mps_available, is_safetensors_available, is_tensorboard_available, is_torch_version, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=False ): '''simple docstring''' try: lowercase = os.environ[key] except KeyError: # KEY isn't set, default to `default`. lowercase = default else: # KEY is set, convert it to True or False. try: lowercase = strtobool(lowerCAmelCase__ ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(f'If set, {key} must be yes or no.' ) return _value lowercase__ :Dict = parse_flag_from_env("RUN_SLOW", default=False) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skip('''Test was skipped''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(_run_slow_tests , '''test is slow''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(not torch.cuda.is_available() , '''test requires only a CPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.cuda.is_available() , '''test requires a GPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_xpu_available() , '''test requires a XPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_mps_available() , '''test requires a `mps` backend support in `torch`''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless( is_transformers_available() and is_datasets_available() , '''test requires the Hugging Face suite''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_bnb_available() , '''test requires the bitsandbytes library''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_tpu_available() , '''test requires TPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.cuda.device_count() == 1 , '''test requires a GPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.xpu.device_count() == 1 , '''test requires a XPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.cuda.device_count() > 1 , '''test requires multiple GPUs''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.xpu.device_count() > 1 , '''test requires multiple XPUs''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_safetensors_available() , '''test requires safetensors''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_deepspeed_available() , '''test requires DeepSpeed''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_torch_version('''>=''' , '''1.12.0''' ) , '''test requires torch version >= 1.12.0''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__=None , lowerCAmelCase__=None ): '''simple docstring''' if test_case is None: return partial(lowerCAmelCase__ , version=lowerCAmelCase__ ) return unittest.skipUnless(is_torch_version('''>=''' , lowerCAmelCase__ ) , f'test requires torch version >= {version}' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_tensorboard_available() , '''test requires Tensorboard''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_wandb_available() , '''test requires wandb''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_comet_ml_available() , '''test requires comet_ml''' )(lowerCAmelCase__ ) lowercase__ :Dict = ( any([is_wandb_available(), is_tensorboard_available()]) and not is_comet_ml_available() ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless( _atleast_one_tracker_available , '''test requires at least one tracker to be available and for `comet_ml` to not be installed''' , )(lowerCAmelCase__ ) class lowercase ( unittest.TestCase ): lowercase_ : int =True @classmethod def A__ ( cls): lowercase = tempfile.mkdtemp() @classmethod def A__ ( cls): if os.path.exists(cls.tmpdir): shutil.rmtree(cls.tmpdir) def A__ ( self): if self.clear_on_setup: for path in Path(self.tmpdir).glob('''**/*'''): if path.is_file(): path.unlink() elif path.is_dir(): shutil.rmtree(A__) class lowercase ( unittest.TestCase ): def A__ ( self): super().tearDown() # Reset the state of the AcceleratorState singleton. AcceleratorState._reset_state() PartialState._reset_state() class lowercase ( unittest.TestCase ): def A__ ( self ,A__): lowercase = mocks if isinstance(A__ ,(tuple, list)) else [mocks] for m in self.mocks: m.start() self.addCleanup(m.stop) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = AcceleratorState() lowercase = tensor[None].clone().to(state.device ) lowercase = gather(lowerCAmelCase__ ).cpu() lowercase = tensor[0].cpu() for i in range(tensors.shape[0] ): if not torch.equal(tensors[i] , lowerCAmelCase__ ): return False return True class lowercase : def __init__( self ,A__ ,A__ ,A__): lowercase = returncode lowercase = stdout lowercase = stderr async def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' while True: lowercase = await stream.readline() if line: callback(lowerCAmelCase__ ) else: break async def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=False , lowerCAmelCase__=False ): '''simple docstring''' if echo: print('''\nRunning: ''' , ''' '''.join(lowerCAmelCase__ ) ) lowercase = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=lowerCAmelCase__ , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=lowerCAmelCase__ , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) lowercase = [] lowercase = [] def tee(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__="" ): lowercase = line.decode('''utf-8''' ).rstrip() sink.append(lowerCAmelCase__ ) if not quiet: print(lowerCAmelCase__ , lowerCAmelCase__ , file=lowerCAmelCase__ ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ asyncio.create_task(_read_stream(p.stdout , lambda lowerCAmelCase__ : tee(lowerCAmelCase__ , lowerCAmelCase__ , sys.stdout , label='''stdout:''' ) ) ), asyncio.create_task(_read_stream(p.stderr , lambda lowerCAmelCase__ : tee(lowerCAmelCase__ , lowerCAmelCase__ , sys.stderr , label='''stderr:''' ) ) ), ] , timeout=lowerCAmelCase__ , ) return _RunOutput(await p.wait() , lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=180 , lowerCAmelCase__=False , lowerCAmelCase__=True ): '''simple docstring''' lowercase = asyncio.get_event_loop() lowercase = loop.run_until_complete( _stream_subprocess(lowerCAmelCase__ , env=lowerCAmelCase__ , stdin=lowerCAmelCase__ , timeout=lowerCAmelCase__ , quiet=lowerCAmelCase__ , echo=lowerCAmelCase__ ) ) lowercase = ''' '''.join(lowerCAmelCase__ ) if result.returncode > 0: lowercase = '''\n'''.join(result.stderr ) raise RuntimeError( f'\'{cmd_str}\' failed with returncode {result.returncode}\n\n' f'The combined stderr from workers follows:\n{stderr}' ) return result class lowercase ( SCREAMING_SNAKE_CASE__ ): pass def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=False ): '''simple docstring''' try: lowercase = subprocess.check_output(lowerCAmelCase__ , stderr=subprocess.STDOUT ) if return_stdout: if hasattr(lowerCAmelCase__ , '''decode''' ): lowercase = output.decode('''utf-8''' ) return output except subprocess.CalledProcessError as e: raise SubprocessCallException( f'Command `{" ".join(lowerCAmelCase__ )}` failed with the following error:\n\n{e.output.decode()}' ) from e
633
1
from __future__ import annotations def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = None ): '''simple docstring''' if start is None: lowercase = 0 if end is None: lowercase = len(lowerCAmelCase__ ) - 1 if start >= end: return lowercase = (start + end) // 2 slowsort(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) slowsort(lowerCAmelCase__ , mid + 1 , lowerCAmelCase__ ) if sequence[end] < sequence[mid]: lowercase , lowercase = sequence[mid], sequence[end] slowsort(lowerCAmelCase__ , lowerCAmelCase__ , end - 1 ) if __name__ == "__main__": from doctest import testmod testmod()
633
import math import time from typing import Dict, List, Optional from torch.utils.data import Dataset from transformers import SeqaSeqTrainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class lowercase ( SCREAMING_SNAKE_CASE__ ): def __init__( self ,*A__ ,A__=None ,A__=None ,**A__): super().__init__(*A__ ,**A__) lowercase = eval_examples lowercase = post_process_function def A__ ( self ,A__ = None ,A__=None ,A__ = None ,A__ = "eval" ,**A__ ,): lowercase = gen_kwargs.copy() lowercase = ( gen_kwargs['''max_length'''] if gen_kwargs.get('''max_length''') is not None else self.args.generation_max_length ) lowercase = ( gen_kwargs['''num_beams'''] if gen_kwargs.get('''num_beams''') is not None else self.args.generation_num_beams ) lowercase = gen_kwargs lowercase = self.eval_dataset if eval_dataset is None else eval_dataset lowercase = self.get_eval_dataloader(A__) lowercase = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. lowercase = self.compute_metrics lowercase = None lowercase = time.time() lowercase = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: lowercase = eval_loop( A__ ,description='''Evaluation''' ,prediction_loss_only=True if compute_metrics is None else None ,ignore_keys=A__ ,metric_key_prefix=A__ ,) finally: lowercase = compute_metrics lowercase = self.args.eval_batch_size * self.args.world_size if f'{metric_key_prefix}_jit_compilation_time' in output.metrics: start_time += output.metrics[f'{metric_key_prefix}_jit_compilation_time'] output.metrics.update( speed_metrics( A__ ,A__ ,num_samples=output.num_samples ,num_steps=math.ceil(output.num_samples / total_batch_size) ,)) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default lowercase = self.post_process_function(A__ ,A__ ,A__) lowercase = self.compute_metrics(A__) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys()): if not key.startswith(f'{metric_key_prefix}_'): lowercase = metrics.pop(A__) metrics.update(output.metrics) else: lowercase = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(A__) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report()) lowercase = self.callback_handler.on_evaluate(self.args ,self.state ,self.control ,A__) return metrics def A__ ( self ,A__ ,A__ ,A__=None ,A__ = "test" ,**A__): lowercase = gen_kwargs.copy() lowercase = self.get_test_dataloader(A__) # Temporarily disable metric computation, we will do it in the loop here. lowercase = self.compute_metrics lowercase = None lowercase = time.time() lowercase = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: lowercase = eval_loop( A__ ,description='''Prediction''' ,prediction_loss_only=True if compute_metrics is None else None ,ignore_keys=A__ ,metric_key_prefix=A__ ,) finally: lowercase = compute_metrics lowercase = self.args.eval_batch_size * self.args.world_size if f'{metric_key_prefix}_jit_compilation_time' in output.metrics: start_time += output.metrics[f'{metric_key_prefix}_jit_compilation_time'] output.metrics.update( speed_metrics( A__ ,A__ ,num_samples=output.num_samples ,num_steps=math.ceil(output.num_samples / total_batch_size) ,)) if self.post_process_function is None or self.compute_metrics is None: return output lowercase = self.post_process_function(A__ ,A__ ,A__ ,'''predict''') lowercase = self.compute_metrics(A__) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys()): if not key.startswith(f'{metric_key_prefix}_'): lowercase = metrics.pop(A__) metrics.update(output.metrics) return PredictionOutput(predictions=predictions.predictions ,label_ids=predictions.label_ids ,metrics=A__)
633
1
from math import acos, sin from typing import List, Tuple, Union import numpy as np import torch from PIL import Image from ...models import AutoencoderKL, UNetaDConditionModel from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import randn_tensor from ..pipeline_utils import AudioPipelineOutput, BaseOutput, DiffusionPipeline, ImagePipelineOutput from .mel import Mel class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Optional[int] =['''vqvae'''] def __init__( self ,A__ ,A__ ,A__ ,A__ ,): super().__init__() self.register_modules(unet=A__ ,scheduler=A__ ,mel=A__ ,vqvae=A__) def A__ ( self): return 5_0 if isinstance(self.scheduler ,A__) else 1_0_0_0 @torch.no_grad() def __call__( self ,A__ = 1 ,A__ = None ,A__ = None ,A__ = 0 ,A__ = 0 ,A__ = None ,A__ = None ,A__ = 0 ,A__ = 0 ,A__ = None ,A__ = 0 ,A__ = None ,A__ = None ,A__=True ,): lowercase = steps or self.get_default_steps() self.scheduler.set_timesteps(A__) lowercase = step_generator or generator # For backwards compatibility if type(self.unet.config.sample_size) == int: lowercase = (self.unet.config.sample_size, self.unet.config.sample_size) if noise is None: lowercase = randn_tensor( ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size[0], self.unet.config.sample_size[1], ) ,generator=A__ ,device=self.device ,) lowercase = noise lowercase = None if audio_file is not None or raw_audio is not None: self.mel.load_audio(A__ ,A__) lowercase = self.mel.audio_slice_to_image(A__) lowercase = np.frombuffer(input_image.tobytes() ,dtype='''uint8''').reshape( (input_image.height, input_image.width)) lowercase = (input_image / 2_5_5) * 2 - 1 lowercase = torch.tensor(input_image[np.newaxis, :, :] ,dtype=torch.float).to(self.device) if self.vqvae is not None: lowercase = self.vqvae.encode(torch.unsqueeze(A__ ,0)).latent_dist.sample( generator=A__)[0] lowercase = self.vqvae.config.scaling_factor * input_images if start_step > 0: lowercase = self.scheduler.add_noise(A__ ,A__ ,self.scheduler.timesteps[start_step - 1]) lowercase = ( self.unet.config.sample_size[1] * self.mel.get_sample_rate() / self.mel.x_res / self.mel.hop_length ) lowercase = int(mask_start_secs * pixels_per_second) lowercase = int(mask_end_secs * pixels_per_second) lowercase = self.scheduler.add_noise(A__ ,A__ ,torch.tensor(self.scheduler.timesteps[start_step:])) for step, t in enumerate(self.progress_bar(self.scheduler.timesteps[start_step:])): if isinstance(self.unet ,A__): lowercase = self.unet(A__ ,A__ ,A__)['''sample'''] else: lowercase = self.unet(A__ ,A__)['''sample'''] if isinstance(self.scheduler ,A__): lowercase = self.scheduler.step( model_output=A__ ,timestep=A__ ,sample=A__ ,eta=A__ ,generator=A__ ,)['''prev_sample'''] else: lowercase = self.scheduler.step( model_output=A__ ,timestep=A__ ,sample=A__ ,generator=A__ ,)['''prev_sample'''] if mask is not None: if mask_start > 0: lowercase = mask[:, step, :, :mask_start] if mask_end > 0: lowercase = mask[:, step, :, -mask_end:] if self.vqvae is not None: # 0.18215 was scaling factor used in training to ensure unit variance lowercase = 1 / self.vqvae.config.scaling_factor * images lowercase = self.vqvae.decode(A__)['''sample'''] lowercase = (images / 2 + 0.5).clamp(0 ,1) lowercase = images.cpu().permute(0 ,2 ,3 ,1).numpy() lowercase = (images * 2_5_5).round().astype('''uint8''') lowercase = list( (Image.fromarray(_[:, :, 0]) for _ in images) if images.shape[3] == 1 else (Image.fromarray(A__ ,mode='''RGB''').convert('''L''') for _ in images)) lowercase = [self.mel.image_to_audio(A__) for _ in images] if not return_dict: return images, (self.mel.get_sample_rate(), audios) return BaseOutput(**AudioPipelineOutput(np.array(A__)[:, np.newaxis, :]) ,**ImagePipelineOutput(A__)) @torch.no_grad() def A__ ( self ,A__ ,A__ = 5_0): assert isinstance(self.scheduler ,A__) self.scheduler.set_timesteps(A__) lowercase = np.array( [np.frombuffer(image.tobytes() ,dtype='''uint8''').reshape((1, image.height, image.width)) for image in images]) lowercase = (sample / 2_5_5) * 2 - 1 lowercase = torch.Tensor(A__).to(self.device) for t in self.progress_bar(torch.flip(self.scheduler.timesteps ,(0,))): lowercase = t - self.scheduler.config.num_train_timesteps // self.scheduler.num_inference_steps lowercase = self.scheduler.alphas_cumprod[t] lowercase = ( self.scheduler.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.scheduler.final_alpha_cumprod ) lowercase = 1 - alpha_prod_t lowercase = self.unet(A__ ,A__)['''sample'''] lowercase = (1 - alpha_prod_t_prev) ** 0.5 * model_output lowercase = (sample - pred_sample_direction) * alpha_prod_t_prev ** (-0.5) lowercase = sample * alpha_prod_t ** 0.5 + beta_prod_t ** 0.5 * model_output return sample @staticmethod def A__ ( A__ ,A__ ,A__): lowercase = acos(torch.dot(torch.flatten(A__) ,torch.flatten(A__)) / torch.norm(A__) / torch.norm(A__)) return sin((1 - alpha) * theta) * xa / sin(A__) + sin(alpha * theta) * xa / sin(A__)
633
# Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import subprocess from packaging.version import Version, parse from accelerate.commands.config.config_args import default_config_file, load_config_from_file lowercase__ :Any = "Run commands across TPU VMs for initial setup before running `accelerate launch`." def UpperCamelCase ( lowerCAmelCase__=None ): '''simple docstring''' if subparsers is not None: lowercase = subparsers.add_parser('''tpu-config''' , description=_description ) else: lowercase = argparse.ArgumentParser('''Accelerate tpu-config command''' , description=_description ) # Core arguments lowercase = parser.add_argument_group( '''Config Arguments''' , '''Arguments that can be configured through `accelerate config`.''' ) config_args.add_argument( '''--config_file''' , type=lowerCAmelCase__ , default=lowerCAmelCase__ , help='''Path to the config file to use for accelerate.''' , ) config_args.add_argument( '''--tpu_name''' , default=lowerCAmelCase__ , help='''The name of the TPU to use. If not specified, will use the TPU specified in the config file.''' , ) config_args.add_argument( '''--tpu_zone''' , default=lowerCAmelCase__ , help='''The zone of the TPU to use. If not specified, will use the zone specified in the config file.''' , ) lowercase = parser.add_argument_group('''TPU Arguments''' , '''Arguments for options ran inside the TPU.''' ) pod_args.add_argument( '''--use_alpha''' , action='''store_true''' , help='''Whether to use `gcloud alpha` when running the TPU training script instead of `gcloud`.''' , ) pod_args.add_argument( '''--command_file''' , default=lowerCAmelCase__ , help='''The path to the file containing the commands to run on the pod on startup.''' , ) pod_args.add_argument( '''--command''' , action='''append''' , nargs='''+''' , help='''A command to run on the pod. Can be passed multiple times.''' , ) pod_args.add_argument( '''--install_accelerate''' , action='''store_true''' , help='''Whether to install accelerate on the pod. Defaults to False.''' , ) pod_args.add_argument( '''--accelerate_version''' , default='''latest''' , help='''The version of accelerate to install on the pod. If not specified, will use the latest pypi version. Specify \'dev\' to install from GitHub.''' , ) pod_args.add_argument( '''--debug''' , action='''store_true''' , help='''If set, will print the command that would be run instead of running it.''' ) if subparsers is not None: parser.set_defaults(func=lowerCAmelCase__ ) return parser def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = None # Get the default from the config file if it exists. if args.config_file is not None or os.path.isfile(lowerCAmelCase__ ): lowercase = load_config_from_file(args.config_file ) if not args.command_file and defaults.command_file is not None and not args.command: lowercase = defaults.command_file if not args.command and defaults.commands is not None: lowercase = defaults.commands if not args.tpu_name: lowercase = defaults.tpu_name if not args.tpu_zone: lowercase = defaults.tpu_zone if args.accelerate_version == "dev": lowercase = '''git+https://github.com/huggingface/accelerate.git''' elif args.accelerate_version == "latest": lowercase = '''accelerate -U''' elif isinstance(parse(args.accelerate_version ) , lowerCAmelCase__ ): lowercase = f'accelerate=={args.accelerate_version}' if not args.command_file and not args.command: raise ValueError('''You must specify either a command file or a command to run on the pod.''' ) if args.command_file: with open(args.command_file , '''r''' ) as f: lowercase = [f.read().splitlines()] # To turn list of lists into list of strings if isinstance(args.command[0] , lowerCAmelCase__ ): lowercase = [line for cmd in args.command for line in cmd] # Default to the shared folder and install accelerate lowercase = ['''cd /usr/share'''] if args.install_accelerate: new_cmd += [f'pip install {args.accelerate_version}'] new_cmd += args.command lowercase = '''; '''.join(lowerCAmelCase__ ) # Then send it to gcloud # Eventually try to use google-api-core to do this instead of subprocess lowercase = ['''gcloud'''] if args.use_alpha: cmd += ["alpha"] cmd += [ "compute", "tpus", "tpu-vm", "ssh", args.tpu_name, "--zone", args.tpu_zone, "--command", args.command, "--worker", "all", ] if args.debug: print(f'Running {" ".join(lowerCAmelCase__ )}' ) return subprocess.run(lowerCAmelCase__ ) print('''Successfully setup pod.''' ) def UpperCamelCase ( ): '''simple docstring''' lowercase = tpu_command_parser() lowercase = parser.parse_args() tpu_command_launcher(lowerCAmelCase__ )
633
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowercase__ :int = { "configuration_roberta": ["ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP", "RobertaConfig", "RobertaOnnxConfig"], "tokenization_roberta": ["RobertaTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :str = ["RobertaTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :List[str] = [ "ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST", "RobertaForCausalLM", "RobertaForMaskedLM", "RobertaForMultipleChoice", "RobertaForQuestionAnswering", "RobertaForSequenceClassification", "RobertaForTokenClassification", "RobertaModel", "RobertaPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :List[Any] = [ "TF_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST", "TFRobertaForCausalLM", "TFRobertaForMaskedLM", "TFRobertaForMultipleChoice", "TFRobertaForQuestionAnswering", "TFRobertaForSequenceClassification", "TFRobertaForTokenClassification", "TFRobertaMainLayer", "TFRobertaModel", "TFRobertaPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :Any = [ "FlaxRobertaForCausalLM", "FlaxRobertaForMaskedLM", "FlaxRobertaForMultipleChoice", "FlaxRobertaForQuestionAnswering", "FlaxRobertaForSequenceClassification", "FlaxRobertaForTokenClassification", "FlaxRobertaModel", "FlaxRobertaPreTrainedModel", ] if TYPE_CHECKING: from .configuration_roberta import ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, RobertaConfig, RobertaOnnxConfig from .tokenization_roberta import RobertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roberta_fast import RobertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roberta import ( ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, RobertaForCausalLM, RobertaForMaskedLM, RobertaForMultipleChoice, RobertaForQuestionAnswering, RobertaForSequenceClassification, RobertaForTokenClassification, RobertaModel, RobertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roberta import ( TF_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForMultipleChoice, TFRobertaForQuestionAnswering, TFRobertaForSequenceClassification, TFRobertaForTokenClassification, TFRobertaMainLayer, TFRobertaModel, TFRobertaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, FlaxRobertaPreTrainedModel, ) else: import sys lowercase__ :Dict = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
633
import json import multiprocessing as mp import re from collections import defaultdict from functools import partial from typing import Dict, List, Optional, Set, Tuple, Type from datasets import Dataset from datasketch import MinHash, MinHashLSH from dpu_utils.utils.iterators import ThreadedIterator from tqdm import tqdm lowercase__ :Optional[Any] = re.compile("[^A-Za-z_0-9]") # parameters used in DuplicationIndex lowercase__ :List[str] = 10 lowercase__ :Tuple = 256 def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if len(lowerCAmelCase__ ) < MIN_NUM_TOKENS: return None lowercase = MinHash(num_perm=lowerCAmelCase__ ) for token in set(lowerCAmelCase__ ): min_hash.update(token.encode() ) return min_hash def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return {t for t in NON_ALPHA.split(lowerCAmelCase__ ) if len(t.strip() ) > 0} class lowercase : def __init__( self ,*, A__ = 0.85 ,): lowercase = duplication_jaccard_threshold lowercase = NUM_PERM lowercase = MinHashLSH(threshold=self._duplication_jaccard_threshold ,num_perm=self._num_perm) lowercase = defaultdict(A__) def A__ ( self ,A__ ,A__): lowercase = self._index.query(A__) if code_key in self._index.keys: print(f'Duplicate key {code_key}') return self._index.insert(A__ ,A__) if len(A__) > 0: for base_duplicate in close_duplicates: if base_duplicate in self._duplicate_clusters: self._duplicate_clusters[base_duplicate].add(A__) break else: self._duplicate_clusters[close_duplicates[0]].add(A__) def A__ ( self): lowercase = [] for base, duplicates in self._duplicate_clusters.items(): lowercase = [base] + list(A__) # reformat the cluster to be a list of dict lowercase = [{'''base_index''': el[0], '''repo_name''': el[1], '''path''': el[2]} for el in cluster] duplicate_clusters.append(A__) return duplicate_clusters def A__ ( self ,A__): lowercase = self.get_duplicate_clusters() with open(A__ ,'''w''') as f: json.dump(A__ ,A__) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase = element lowercase = get_min_hash([t for t in NON_ALPHA.split(data['''content'''] ) if len(t.strip() ) > 0] ) if min_hash is not None: return (index, data["repo_name"], data["path"]), min_hash def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' with mp.Pool() as pool: for data in pool.imap_unordered( _compute_min_hash , ThreadedIterator(lowerCAmelCase__ , max_queue_size=1_0000 ) , chunksize=100 , ): if data is not None: yield data def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = DuplicationIndex(duplication_jaccard_threshold=lowerCAmelCase__ ) for filename, min_hash in tqdm(ThreadedIterator(minhash_iter(enumerate(lowerCAmelCase__ ) ) , max_queue_size=100 ) ): di.add(lowerCAmelCase__ , lowerCAmelCase__ ) # Returns a List[Cluster] where Cluster is List[str] with the filenames. return di.get_duplicate_clusters() def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = get_tokens(lowerCAmelCase__ ) lowercase = get_tokens(lowerCAmelCase__ ) return len(tokensa & tokensa ) / len(tokensa | tokensa ) lowercase__ :List[Any] = None def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = [] for elementa in cluster: lowercase = _shared_dataset[elementa['''base_index''']]['''content'''] for elementa in extremes: lowercase = _shared_dataset[elementa['''base_index''']]['''content'''] if jaccard_similarity(lowerCAmelCase__ , lowerCAmelCase__ ) >= jaccard_threshold: elementa["copies"] += 1 break else: lowercase = 1 extremes.append(lowerCAmelCase__ ) return extremes def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' global _shared_dataset lowercase = dataset lowercase = [] lowercase = partial(_find_cluster_extremes_shared , jaccard_threshold=lowerCAmelCase__ ) with mp.Pool() as pool: for extremes in tqdm( pool.imap_unordered( lowerCAmelCase__ , lowerCAmelCase__ , ) , total=len(lowerCAmelCase__ ) , ): extremes_list.append(lowerCAmelCase__ ) return extremes_list def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ = 0.85 ): '''simple docstring''' lowercase = make_duplicate_clusters(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase = {x['''base_index'''] for cluster in duplicate_clusters for x in cluster} lowercase = {} lowercase = find_extremes(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) for extremes in extremes_clusters: for element in extremes: lowercase = element lowercase = duplicate_indices - set(extreme_dict.keys() ) lowercase = dataset.filter(lambda lowerCAmelCase__ , lowerCAmelCase__ : idx not in remove_indices , with_indices=lowerCAmelCase__ ) # update duplicate_clusters for cluster in duplicate_clusters: for element in cluster: lowercase = element['''base_index'''] in extreme_dict if element["is_extreme"]: lowercase = extreme_dict[element['''base_index''']]['''copies'''] print(f'Original dataset size: {len(lowerCAmelCase__ )}' ) print(f'Number of duplicate clusters: {len(lowerCAmelCase__ )}' ) print(f'Files in duplicate cluster: {len(lowerCAmelCase__ )}' ) print(f'Unique files in duplicate cluster: {len(lowerCAmelCase__ )}' ) print(f'Filtered dataset size: {len(lowerCAmelCase__ )}' ) return ds_filter, duplicate_clusters
633
1
from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ :Union[str, Any] = logging.get_logger(__name__) lowercase__ :int = { "microsoft/biogpt": "https://huggingface.co/microsoft/biogpt/resolve/main/config.json", # See all BioGPT models at https://huggingface.co/models?filter=biogpt } class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : int ='''biogpt''' def __init__( self ,A__=4_2_3_8_4 ,A__=1_0_2_4 ,A__=2_4 ,A__=1_6 ,A__=4_0_9_6 ,A__="gelu" ,A__=0.1 ,A__=0.1 ,A__=1_0_2_4 ,A__=0.02 ,A__=1E-12 ,A__=True ,A__=True ,A__=0.0 ,A__=0.0 ,A__=1 ,A__=0 ,A__=2 ,**A__ ,): lowercase = vocab_size lowercase = max_position_embeddings lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_attention_heads lowercase = intermediate_size lowercase = hidden_act lowercase = hidden_dropout_prob lowercase = attention_probs_dropout_prob lowercase = initializer_range lowercase = layer_norm_eps lowercase = scale_embedding lowercase = use_cache lowercase = layerdrop lowercase = activation_dropout super().__init__(pad_token_id=A__ ,bos_token_id=A__ ,eos_token_id=A__ ,**A__)
633
import numpy as np import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModelWithProjection, PreTrainedModel from ...utils import logging lowercase__ :Union[str, Any] = logging.get_logger(__name__) class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Union[str, Any] =CLIPConfig lowercase_ : str =['''CLIPEncoderLayer'''] def __init__( self ,A__): super().__init__(A__) lowercase = CLIPVisionModelWithProjection(config.vision_config) lowercase = nn.Linear(config.vision_config.projection_dim ,1) lowercase = nn.Linear(config.vision_config.projection_dim ,1) @torch.no_grad() def A__ ( self ,A__ ,A__ ,A__=0.5 ,A__=0.5): lowercase = self.vision_model(A__)[0] lowercase = self.p_head(A__) lowercase = nsfw_detected.flatten() lowercase = nsfw_detected > p_threshold lowercase = nsfw_detected.tolist() if any(A__): logger.warning( '''Potential NSFW content was detected in one or more images. A black image will be returned instead.''' ''' Try again with a different prompt and/or seed.''') for idx, nsfw_detected_ in enumerate(A__): if nsfw_detected_: lowercase = np.zeros(images[idx].shape) lowercase = self.w_head(A__) lowercase = watermark_detected.flatten() lowercase = watermark_detected > w_threshold lowercase = watermark_detected.tolist() if any(A__): logger.warning( '''Potential watermarked content was detected in one or more images. A black image will be returned instead.''' ''' Try again with a different prompt and/or seed.''') for idx, watermark_detected_ in enumerate(A__): if watermark_detected_: lowercase = np.zeros(images[idx].shape) return images, nsfw_detected, watermark_detected
633
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) lowercase__ :Optional[int] = { "configuration_owlvit": [ "OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "OwlViTConfig", "OwlViTOnnxConfig", "OwlViTTextConfig", "OwlViTVisionConfig", ], "processing_owlvit": ["OwlViTProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :Tuple = ["OwlViTFeatureExtractor"] lowercase__ :Union[str, Any] = ["OwlViTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :Dict = [ "OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "OwlViTModel", "OwlViTPreTrainedModel", "OwlViTTextModel", "OwlViTVisionModel", "OwlViTForObjectDetection", ] if TYPE_CHECKING: from .configuration_owlvit import ( OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, OwlViTConfig, OwlViTOnnxConfig, OwlViTTextConfig, OwlViTVisionConfig, ) from .processing_owlvit import OwlViTProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_owlvit import OwlViTFeatureExtractor from .image_processing_owlvit import OwlViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_owlvit import ( OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST, OwlViTForObjectDetection, OwlViTModel, OwlViTPreTrainedModel, OwlViTTextModel, OwlViTVisionModel, ) else: import sys lowercase__ :Dict = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
633
class lowercase : def __init__( self ,A__): lowercase = val lowercase = None lowercase = None def A__ ( self ,A__): if self.val: if val < self.val: if self.left is None: lowercase = Node(A__) else: self.left.insert(A__) elif val > self.val: if self.right is None: lowercase = Node(A__) else: self.right.insert(A__) else: lowercase = val def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' # Recursive traversal if root: inorder(root.left , lowerCAmelCase__ ) res.append(root.val ) inorder(root.right , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' # Build BST if len(lowerCAmelCase__ ) == 0: return arr lowercase = Node(arr[0] ) for i in range(1 , len(lowerCAmelCase__ ) ): root.insert(arr[i] ) # Traverse BST in order. lowercase = [] inorder(lowerCAmelCase__ , lowerCAmelCase__ ) return res if __name__ == "__main__": print(tree_sort([10, 1, 3, 2, 9, 14, 13]))
633
1
def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' return int(input_a == input_a == 0 ) def UpperCamelCase ( ): '''simple docstring''' print('''Truth Table of NOR Gate:''' ) print('''| Input 1 | Input 2 | Output |''' ) print(f'| 0 | 0 | {nor_gate(0 , 0 )} |' ) print(f'| 0 | 1 | {nor_gate(0 , 1 )} |' ) print(f'| 1 | 0 | {nor_gate(1 , 0 )} |' ) print(f'| 1 | 1 | {nor_gate(1 , 1 )} |' ) if __name__ == "__main__": import doctest doctest.testmod() main()
633
import os def UpperCamelCase ( lowerCAmelCase__ = "input.txt" ): '''simple docstring''' with open(os.path.join(os.path.dirname(lowerCAmelCase__ ) , lowerCAmelCase__ ) ) as input_file: lowercase = [ [int(lowerCAmelCase__ ) for element in line.split(''',''' )] for line in input_file.readlines() ] lowercase = len(lowerCAmelCase__ ) lowercase = len(matrix[0] ) lowercase = [[-1 for _ in range(lowerCAmelCase__ )] for _ in range(lowerCAmelCase__ )] for i in range(lowerCAmelCase__ ): lowercase = matrix[i][0] for j in range(1 , lowerCAmelCase__ ): for i in range(lowerCAmelCase__ ): lowercase = minimal_path_sums[i][j - 1] + matrix[i][j] for i in range(1 , lowerCAmelCase__ ): lowercase = min( minimal_path_sums[i][j] , minimal_path_sums[i - 1][j] + matrix[i][j] ) for i in range(rows - 2 , -1 , -1 ): lowercase = min( minimal_path_sums[i][j] , minimal_path_sums[i + 1][j] + matrix[i][j] ) return min(minimal_path_sums_row[-1] for minimal_path_sums_row in minimal_path_sums ) if __name__ == "__main__": print(F'{solution() = }')
633
1
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetrImageProcessor class lowercase ( unittest.TestCase ): def __init__( self ,A__ ,A__=7 ,A__=3 ,A__=3_0 ,A__=4_0_0 ,A__=True ,A__=None ,A__=True ,A__=1 / 2_5_5 ,A__=True ,A__=[0.5, 0.5, 0.5] ,A__=[0.5, 0.5, 0.5] ,A__=True ,): # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p lowercase = size if size is not None else {'''shortest_edge''': 1_8, '''longest_edge''': 1_3_3_3} lowercase = parent lowercase = batch_size lowercase = num_channels lowercase = min_resolution lowercase = max_resolution lowercase = do_resize lowercase = size lowercase = do_rescale lowercase = rescale_factor lowercase = do_normalize lowercase = image_mean lowercase = image_std lowercase = do_pad def A__ ( self): return { "do_resize": self.do_resize, "size": self.size, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_pad": self.do_pad, } def A__ ( self ,A__ ,A__=False): if not batched: lowercase = image_inputs[0] if isinstance(A__ ,Image.Image): lowercase , lowercase = image.size else: lowercase , lowercase = image.shape[1], image.shape[2] if w < h: lowercase = int(self.size['''shortest_edge'''] * h / w) lowercase = self.size['''shortest_edge'''] elif w > h: lowercase = self.size['''shortest_edge'''] lowercase = int(self.size['''shortest_edge'''] * w / h) else: lowercase = self.size['''shortest_edge'''] lowercase = self.size['''shortest_edge'''] else: lowercase = [] for image in image_inputs: lowercase , lowercase = self.get_expected_values([image]) expected_values.append((expected_height, expected_width)) lowercase = max(A__ ,key=lambda A__: item[0])[0] lowercase = max(A__ ,key=lambda A__: item[1])[1] return expected_height, expected_width @require_torch @require_vision class lowercase ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): lowercase_ : Tuple =DetrImageProcessor if is_vision_available() else None def A__ ( self): lowercase = DetrImageProcessingTester(self) @property def A__ ( self): return self.image_processor_tester.prepare_image_processor_dict() def A__ ( self): lowercase = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(A__ ,'''image_mean''')) self.assertTrue(hasattr(A__ ,'''image_std''')) self.assertTrue(hasattr(A__ ,'''do_normalize''')) self.assertTrue(hasattr(A__ ,'''do_rescale''')) self.assertTrue(hasattr(A__ ,'''rescale_factor''')) self.assertTrue(hasattr(A__ ,'''do_resize''')) self.assertTrue(hasattr(A__ ,'''size''')) self.assertTrue(hasattr(A__ ,'''do_pad''')) def A__ ( self): lowercase = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size ,{'''shortest_edge''': 1_8, '''longest_edge''': 1_3_3_3}) self.assertEqual(image_processor.do_pad ,A__) lowercase = self.image_processing_class.from_dict( self.image_processor_dict ,size=4_2 ,max_size=8_4 ,pad_and_return_pixel_mask=A__) self.assertEqual(image_processor.size ,{'''shortest_edge''': 4_2, '''longest_edge''': 8_4}) self.assertEqual(image_processor.do_pad ,A__) def A__ ( self): pass def A__ ( self): # Initialize image_processing lowercase = self.image_processing_class(**self.image_processor_dict) # create random PIL images lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=A__) for image in image_inputs: self.assertIsInstance(A__ ,Image.Image) # Test not batched input lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''').pixel_values lowercase , lowercase = self.image_processor_tester.get_expected_values(A__) self.assertEqual( encoded_images.shape ,(1, self.image_processor_tester.num_channels, expected_height, expected_width) ,) # Test batched lowercase , lowercase = self.image_processor_tester.get_expected_values(A__ ,batched=A__) lowercase = image_processing(A__ ,return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) ,) def A__ ( self): # Initialize image_processing lowercase = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=A__ ,numpify=A__) for image in image_inputs: self.assertIsInstance(A__ ,np.ndarray) # Test not batched input lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''').pixel_values lowercase , lowercase = self.image_processor_tester.get_expected_values(A__) self.assertEqual( encoded_images.shape ,(1, self.image_processor_tester.num_channels, expected_height, expected_width) ,) # Test batched lowercase = image_processing(A__ ,return_tensors='''pt''').pixel_values lowercase , lowercase = self.image_processor_tester.get_expected_values(A__ ,batched=A__) self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) ,) def A__ ( self): # Initialize image_processing lowercase = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=A__ ,torchify=A__) for image in image_inputs: self.assertIsInstance(A__ ,torch.Tensor) # Test not batched input lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''').pixel_values lowercase , lowercase = self.image_processor_tester.get_expected_values(A__) self.assertEqual( encoded_images.shape ,(1, self.image_processor_tester.num_channels, expected_height, expected_width) ,) # Test batched lowercase = image_processing(A__ ,return_tensors='''pt''').pixel_values lowercase , lowercase = self.image_processor_tester.get_expected_values(A__ ,batched=A__) self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) ,) @slow def A__ ( self): # prepare image and target lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''') with open('''./tests/fixtures/tests_samples/COCO/coco_annotations.txt''' ,'''r''') as f: lowercase = json.loads(f.read()) lowercase = {'''image_id''': 3_9_7_6_9, '''annotations''': target} # encode them lowercase = DetrImageProcessor.from_pretrained('''facebook/detr-resnet-50''') lowercase = image_processing(images=A__ ,annotations=A__ ,return_tensors='''pt''') # verify pixel values lowercase = torch.Size([1, 3, 8_0_0, 1_0_6_6]) self.assertEqual(encoding['''pixel_values'''].shape ,A__) lowercase = torch.tensor([0.2796, 0.3138, 0.3481]) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] ,A__ ,atol=1E-4)) # verify area lowercase = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] ,A__)) # verify boxes lowercase = torch.Size([6, 4]) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape ,A__) lowercase = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] ,A__ ,atol=1E-3)) # verify image_id lowercase = torch.tensor([3_9_7_6_9]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] ,A__)) # verify is_crowd lowercase = torch.tensor([0, 0, 0, 0, 0, 0]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] ,A__)) # verify class_labels lowercase = torch.tensor([7_5, 7_5, 6_3, 6_5, 1_7, 1_7]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] ,A__)) # verify orig_size lowercase = torch.tensor([4_8_0, 6_4_0]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] ,A__)) # verify size lowercase = torch.tensor([8_0_0, 1_0_6_6]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] ,A__)) @slow def A__ ( self): # prepare image, target and masks_path lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''') with open('''./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt''' ,'''r''') as f: lowercase = json.loads(f.read()) lowercase = {'''file_name''': '''000000039769.png''', '''image_id''': 3_9_7_6_9, '''segments_info''': target} lowercase = pathlib.Path('''./tests/fixtures/tests_samples/COCO/coco_panoptic''') # encode them lowercase = DetrImageProcessor.from_pretrained('''facebook/detr-resnet-50-panoptic''') lowercase = image_processing(images=A__ ,annotations=A__ ,masks_path=A__ ,return_tensors='''pt''') # verify pixel values lowercase = torch.Size([1, 3, 8_0_0, 1_0_6_6]) self.assertEqual(encoding['''pixel_values'''].shape ,A__) lowercase = torch.tensor([0.2796, 0.3138, 0.3481]) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] ,A__ ,atol=1E-4)) # verify area lowercase = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] ,A__)) # verify boxes lowercase = torch.Size([6, 4]) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape ,A__) lowercase = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] ,A__ ,atol=1E-3)) # verify image_id lowercase = torch.tensor([3_9_7_6_9]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] ,A__)) # verify is_crowd lowercase = torch.tensor([0, 0, 0, 0, 0, 0]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] ,A__)) # verify class_labels lowercase = torch.tensor([1_7, 1_7, 6_3, 7_5, 7_5, 9_3]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] ,A__)) # verify masks lowercase = 8_2_2_8_7_3 self.assertEqual(encoding['''labels'''][0]['''masks'''].sum().item() ,A__) # verify orig_size lowercase = torch.tensor([4_8_0, 6_4_0]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] ,A__)) # verify size lowercase = torch.tensor([8_0_0, 1_0_6_6]) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] ,A__))
633
from __future__ import annotations def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if len(lowerCAmelCase__ ) < 2: raise ValueError('''Monogons and Digons are not polygons in the Euclidean space''' ) if any(i <= 0 for i in nums ): raise ValueError('''All values must be greater than 0''' ) lowercase = nums.copy() copy_nums.sort() return copy_nums[-1] < sum(copy_nums[:-1] ) if __name__ == "__main__": import doctest doctest.testmod()
633
1
import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.data import Dataset from transformers import BartTokenizer, RagTokenizer, TaTokenizer def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=True , lowerCAmelCase__="pt" ): '''simple docstring''' lowercase = {'''add_prefix_space''': True} if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) and not line.startswith(''' ''' ) else {} lowercase = padding_side return tokenizer( [line] , max_length=lowerCAmelCase__ , padding='''max_length''' if pad_to_max_length else None , truncation=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , **lowerCAmelCase__ , ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None , ): '''simple docstring''' lowercase = input_ids.ne(lowerCAmelCase__ ).any(dim=0 ) if attention_mask is None: return input_ids[:, keep_column_mask] else: return (input_ids[:, keep_column_mask], attention_mask[:, keep_column_mask]) class lowercase ( SCREAMING_SNAKE_CASE__ ): def __init__( self ,A__ ,A__ ,A__ ,A__ ,A__="train" ,A__=None ,A__=None ,A__=None ,A__="" ,): super().__init__() lowercase = Path(A__).joinpath(type_path + '''.source''') lowercase = Path(A__).joinpath(type_path + '''.target''') lowercase = self.get_char_lens(self.src_file) lowercase = max_source_length lowercase = max_target_length assert min(self.src_lens) > 0, f'found empty line in {self.src_file}' lowercase = tokenizer lowercase = prefix if n_obs is not None: lowercase = self.src_lens[:n_obs] lowercase = src_lang lowercase = tgt_lang def __len__( self): return len(self.src_lens) def __getitem__( self ,A__): lowercase = index + 1 # linecache starts at 1 lowercase = self.prefix + linecache.getline(str(self.src_file) ,A__).rstrip('''\n''') lowercase = linecache.getline(str(self.tgt_file) ,A__).rstrip('''\n''') assert source_line, f'empty source line for index {index}' assert tgt_line, f'empty tgt line for index {index}' # Need to add eos token manually for T5 if isinstance(self.tokenizer ,A__): source_line += self.tokenizer.eos_token tgt_line += self.tokenizer.eos_token # Pad source and target to the right lowercase = ( self.tokenizer.question_encoder if isinstance(self.tokenizer ,A__) else self.tokenizer ) lowercase = self.tokenizer.generator if isinstance(self.tokenizer ,A__) else self.tokenizer lowercase = encode_line(A__ ,A__ ,self.max_source_length ,'''right''') lowercase = encode_line(A__ ,A__ ,self.max_target_length ,'''right''') lowercase = source_inputs['''input_ids'''].squeeze() lowercase = target_inputs['''input_ids'''].squeeze() lowercase = source_inputs['''attention_mask'''].squeeze() return { "input_ids": source_ids, "attention_mask": src_mask, "decoder_input_ids": target_ids, } @staticmethod def A__ ( A__): return [len(A__) for x in Path(A__).open().readlines()] def A__ ( self ,A__): lowercase = torch.stack([x['''input_ids'''] for x in batch]) lowercase = torch.stack([x['''attention_mask'''] for x in batch]) lowercase = torch.stack([x['''decoder_input_ids'''] for x in batch]) lowercase = ( self.tokenizer.generator.pad_token_id if isinstance(self.tokenizer ,A__) else self.tokenizer.pad_token_id ) lowercase = ( self.tokenizer.question_encoder.pad_token_id if isinstance(self.tokenizer ,A__) else self.tokenizer.pad_token_id ) lowercase = trim_batch(A__ ,A__) lowercase , lowercase = trim_batch(A__ ,A__ ,attention_mask=A__) lowercase = { '''input_ids''': source_ids, '''attention_mask''': source_mask, '''decoder_input_ids''': y, } return batch lowercase__ :str = getLogger(__name__) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return list(itertools.chain.from_iterable(lowerCAmelCase__ ) ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = get_git_info() save_json(lowerCAmelCase__ , os.path.join(lowerCAmelCase__ , '''git_log.json''' ) ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=4 , **lowerCAmelCase__ ): '''simple docstring''' with open(lowerCAmelCase__ , '''w''' ) as f: json.dump(lowerCAmelCase__ , lowerCAmelCase__ , indent=lowerCAmelCase__ , **lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' with open(lowerCAmelCase__ ) as f: return json.load(lowerCAmelCase__ ) def UpperCamelCase ( ): '''simple docstring''' lowercase = git.Repo(search_parent_directories=lowerCAmelCase__ ) lowercase = { '''repo_id''': str(lowerCAmelCase__ ), '''repo_sha''': str(repo.head.object.hexsha ), '''repo_branch''': str(repo.active_branch ), '''hostname''': str(socket.gethostname() ), } return repo_infos def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' return list(map(lowerCAmelCase__ , lowerCAmelCase__ ) ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' with open(lowerCAmelCase__ , '''wb''' ) as f: return pickle.dump(lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' def remove_articles(lowerCAmelCase__ ): return re.sub(R'''\b(a|an|the)\b''' , ''' ''' , lowerCAmelCase__ ) def white_space_fix(lowerCAmelCase__ ): return " ".join(text.split() ) def remove_punc(lowerCAmelCase__ ): lowercase = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(lowerCAmelCase__ ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(lowerCAmelCase__ ) ) ) ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = normalize_answer(lowerCAmelCase__ ).split() lowercase = normalize_answer(lowerCAmelCase__ ).split() lowercase = Counter(lowerCAmelCase__ ) & Counter(lowerCAmelCase__ ) lowercase = sum(common.values() ) if num_same == 0: return 0 lowercase = 1.0 * num_same / len(lowerCAmelCase__ ) lowercase = 1.0 * num_same / len(lowerCAmelCase__ ) lowercase = (2 * precision * recall) / (precision + recall) return fa def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' return normalize_answer(lowerCAmelCase__ ) == normalize_answer(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' assert len(lowerCAmelCase__ ) == len(lowerCAmelCase__ ) lowercase = 0 for hypo, pred in zip(lowerCAmelCase__ , lowerCAmelCase__ ): em += exact_match_score(lowerCAmelCase__ , lowerCAmelCase__ ) if len(lowerCAmelCase__ ) > 0: em /= len(lowerCAmelCase__ ) return {"em": em} def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return model_prefix.startswith('''rag''' ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = {p: p for p in extra_params} # T5 models don't have `dropout` param, they have `dropout_rate` instead lowercase = '''dropout_rate''' for p in extra_params: if getattr(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): if not hasattr(lowerCAmelCase__ , lowerCAmelCase__ ) and not hasattr(lowerCAmelCase__ , equivalent_param[p] ): logger.info('''config doesn\'t have a `{}` attribute'''.format(lowerCAmelCase__ ) ) delattr(lowerCAmelCase__ , lowerCAmelCase__ ) continue lowercase = p if hasattr(lowerCAmelCase__ , lowerCAmelCase__ ) else equivalent_param[p] setattr(lowerCAmelCase__ , lowerCAmelCase__ , getattr(lowerCAmelCase__ , lowerCAmelCase__ ) ) delattr(lowerCAmelCase__ , lowerCAmelCase__ ) return hparams, config
633
def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' return int((input_a, input_a).count(0 ) == 0 ) def UpperCamelCase ( ): '''simple docstring''' assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_gate(1 , 0 ) == 0 assert and_gate(1 , 1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
633
1
from math import factorial def UpperCamelCase ( lowerCAmelCase__ = 100 ): '''simple docstring''' return sum(int(lowerCAmelCase__ ) for x in str(factorial(lowerCAmelCase__ ) ) ) if __name__ == "__main__": print(solution(int(input("Enter the Number: ").strip())))
633
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowercase__ :Tuple = { "configuration_biogpt": ["BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BioGptConfig"], "tokenization_biogpt": ["BioGptTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :Union[str, Any] = [ "BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST", "BioGptForCausalLM", "BioGptForTokenClassification", "BioGptForSequenceClassification", "BioGptModel", "BioGptPreTrainedModel", ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys lowercase__ :Any = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
633
1
import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py lowercase__ :int = "src/diffusers" lowercase__ :List[Any] = "." # This is to make sure the diffusers module imported is the one in the repo. lowercase__ :Optional[Any] = importlib.util.spec_from_file_location( "diffusers", os.path.join(DIFFUSERS_PATH, "__init__.py"), submodule_search_locations=[DIFFUSERS_PATH], ) lowercase__ :Union[str, Any] = spec.loader.load_module() def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' return line.startswith(lowerCAmelCase__ ) or len(lowerCAmelCase__ ) <= 1 or re.search(R'''^\s*\)(\s*->.*:|:)\s*$''' , lowerCAmelCase__ ) is not None def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = object_name.split('''.''' ) lowercase = 0 # First let's find the module where our object lives. lowercase = parts[i] while i < len(lowerCAmelCase__ ) and not os.path.isfile(os.path.join(lowerCAmelCase__ , f'{module}.py' ) ): i += 1 if i < len(lowerCAmelCase__ ): lowercase = os.path.join(lowerCAmelCase__ , parts[i] ) if i >= len(lowerCAmelCase__ ): raise ValueError(f'`object_name` should begin with the name of a module of diffusers but got {object_name}.' ) with open(os.path.join(lowerCAmelCase__ , f'{module}.py' ) , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowercase = f.readlines() # Now let's find the class / func in the code! lowercase = '''''' lowercase = 0 for name in parts[i + 1 :]: while ( line_index < len(lowerCAmelCase__ ) and re.search(Rf'^{indent}(class|def)\s+{name}(\(|\:)' , lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(lowerCAmelCase__ ): raise ValueError(f' {object_name} does not match any function or class in {module}.' ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). lowercase = line_index while line_index < len(lowerCAmelCase__ ) and _should_continue(lines[line_index] , lowerCAmelCase__ ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 lowercase = lines[start_index:line_index] return "".join(lowerCAmelCase__ ) lowercase__ :Dict = re.compile(r"^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)") lowercase__ :Union[str, Any] = re.compile(r"^\s*(\S+)->(\S+)(\s+.*|$)") lowercase__ :Dict = re.compile(r"<FILL\s+[^>]*>") def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = code.split('''\n''' ) lowercase = 0 while idx < len(lowerCAmelCase__ ) and len(lines[idx] ) == 0: idx += 1 if idx < len(lowerCAmelCase__ ): return re.search(R'''^(\s*)\S''' , lines[idx] ).groups()[0] return "" def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = len(get_indent(lowerCAmelCase__ ) ) > 0 if has_indent: lowercase = f'class Bla:\n{code}' lowercase = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 , preview=lowerCAmelCase__ ) lowercase = black.format_str(lowerCAmelCase__ , mode=lowerCAmelCase__ ) lowercase , lowercase = style_docstrings_in_code(lowerCAmelCase__ ) return result[len('''class Bla:\n''' ) :] if has_indent else result def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=False ): '''simple docstring''' with open(lowerCAmelCase__ , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowercase = f.readlines() lowercase = [] lowercase = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(lowerCAmelCase__ ): lowercase = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. lowercase , lowercase , lowercase = search.groups() lowercase = find_code_in_diffusers(lowerCAmelCase__ ) lowercase = get_indent(lowerCAmelCase__ ) lowercase = line_index + 1 if indent == theoretical_indent else line_index + 2 lowercase = theoretical_indent lowercase = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. lowercase = True while line_index < len(lowerCAmelCase__ ) and should_continue: line_index += 1 if line_index >= len(lowerCAmelCase__ ): break lowercase = lines[line_index] lowercase = _should_continue(lowerCAmelCase__ , lowerCAmelCase__ ) and re.search(f'^{indent}# End copy' , lowerCAmelCase__ ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 lowercase = lines[start_index:line_index] lowercase = ''''''.join(lowerCAmelCase__ ) # Remove any nested `Copied from` comments to avoid circular copies lowercase = [line for line in theoretical_code.split('''\n''' ) if _re_copy_warning.search(lowerCAmelCase__ ) is None] lowercase = '''\n'''.join(lowerCAmelCase__ ) # Before comparing, use the `replace_pattern` on the original code. if len(lowerCAmelCase__ ) > 0: lowercase = replace_pattern.replace('''with''' , '''''' ).split(''',''' ) lowercase = [_re_replace_pattern.search(lowerCAmelCase__ ) for p in patterns] for pattern in patterns: if pattern is None: continue lowercase , lowercase , lowercase = pattern.groups() lowercase = re.sub(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) if option.strip() == "all-casing": lowercase = re.sub(obja.lower() , obja.lower() , lowerCAmelCase__ ) lowercase = re.sub(obja.upper() , obja.upper() , lowerCAmelCase__ ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line lowercase = blackify(lines[start_index - 1] + theoretical_code ) lowercase = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: lowercase = lines[:start_index] + [theoretical_code] + lines[line_index:] lowercase = start_index + 1 if overwrite and len(lowerCAmelCase__ ) > 0: # Warn the user a file has been modified. print(f'Detected changes, rewriting {filename}.' ) with open(lowerCAmelCase__ , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(lowerCAmelCase__ ) return diffs def UpperCamelCase ( lowerCAmelCase__ = False ): '''simple docstring''' lowercase = glob.glob(os.path.join(lowerCAmelCase__ , '''**/*.py''' ) , recursive=lowerCAmelCase__ ) lowercase = [] for filename in all_files: lowercase = is_copy_consistent(lowerCAmelCase__ , lowerCAmelCase__ ) diffs += [f'- {filename}: copy does not match {d[0]} at line {d[1]}' for d in new_diffs] if not overwrite and len(lowerCAmelCase__ ) > 0: lowercase = '''\n'''.join(lowerCAmelCase__ ) raise Exception( '''Found the following copy inconsistencies:\n''' + diff + '''\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.''' ) if __name__ == "__main__": lowercase__ :int = argparse.ArgumentParser() parser.add_argument("--fix_and_overwrite", action="store_true", help="Whether to fix inconsistencies.") lowercase__ :Optional[Any] = parser.parse_args() check_copies(args.fix_and_overwrite)
633
import logging from transformers import PretrainedConfig lowercase__ :int = logging.getLogger(__name__) lowercase__ :Dict = { "bertabs-finetuned-cnndm": "https://huggingface.co/remi/bertabs-finetuned-cnndm-extractive-abstractive-summarization/resolve/main/config.json", } class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Optional[int] ='''bertabs''' def __init__( self ,A__=3_0_5_2_2 ,A__=5_1_2 ,A__=6 ,A__=5_1_2 ,A__=8 ,A__=5_1_2 ,A__=0.2 ,A__=6 ,A__=7_6_8 ,A__=8 ,A__=2_0_4_8 ,A__=0.2 ,**A__ ,): super().__init__(**A__) lowercase = vocab_size lowercase = max_pos lowercase = enc_layers lowercase = enc_hidden_size lowercase = enc_heads lowercase = enc_ff_size lowercase = enc_dropout lowercase = dec_layers lowercase = dec_hidden_size lowercase = dec_heads lowercase = dec_ff_size lowercase = dec_dropout
633
1
from typing import Any, Dict, Optional import torch import torch.nn.functional as F from torch import nn from ..utils import maybe_allow_in_graph from .activations import get_activation from .attention_processor import Attention from .embeddings import CombinedTimestepLabelEmbeddings @maybe_allow_in_graph class lowercase ( nn.Module ): def __init__( self ,A__ ,A__ ,A__ ,A__=0.0 ,A__ = None ,A__ = "geglu" ,A__ = None ,A__ = False ,A__ = False ,A__ = False ,A__ = False ,A__ = True ,A__ = "layer_norm" ,A__ = False ,): super().__init__() lowercase = only_cross_attention lowercase = (num_embeds_ada_norm is not None) and norm_type == '''ada_norm_zero''' lowercase = (num_embeds_ada_norm is not None) and norm_type == '''ada_norm''' if norm_type in ("ada_norm", "ada_norm_zero") and num_embeds_ada_norm is None: raise ValueError( f'`norm_type` is set to {norm_type}, but `num_embeds_ada_norm` is not defined. Please make sure to' f' define `num_embeds_ada_norm` if setting `norm_type` to {norm_type}.') # Define 3 blocks. Each block has its own normalization layer. # 1. Self-Attn if self.use_ada_layer_norm: lowercase = AdaLayerNorm(A__ ,A__) elif self.use_ada_layer_norm_zero: lowercase = AdaLayerNormZero(A__ ,A__) else: lowercase = nn.LayerNorm(A__ ,elementwise_affine=A__) lowercase = Attention( query_dim=A__ ,heads=A__ ,dim_head=A__ ,dropout=A__ ,bias=A__ ,cross_attention_dim=cross_attention_dim if only_cross_attention else None ,upcast_attention=A__ ,) # 2. Cross-Attn if cross_attention_dim is not None or double_self_attention: # We currently only use AdaLayerNormZero for self attention where there will only be one attention block. # I.e. the number of returned modulation chunks from AdaLayerZero would not make sense if returned during # the second cross attention block. lowercase = ( AdaLayerNorm(A__ ,A__) if self.use_ada_layer_norm else nn.LayerNorm(A__ ,elementwise_affine=A__) ) lowercase = Attention( query_dim=A__ ,cross_attention_dim=cross_attention_dim if not double_self_attention else None ,heads=A__ ,dim_head=A__ ,dropout=A__ ,bias=A__ ,upcast_attention=A__ ,) # is self-attn if encoder_hidden_states is none else: lowercase = None lowercase = None # 3. Feed-forward lowercase = nn.LayerNorm(A__ ,elementwise_affine=A__) lowercase = FeedForward(A__ ,dropout=A__ ,activation_fn=A__ ,final_dropout=A__) # let chunk size default to None lowercase = None lowercase = 0 def A__ ( self ,A__ ,A__): # Sets chunk feed-forward lowercase = chunk_size lowercase = dim def A__ ( self ,A__ ,A__ = None ,A__ = None ,A__ = None ,A__ = None ,A__ = None ,A__ = None ,): # Notice that normalization is always applied before the real computation in the following blocks. # 1. Self-Attention if self.use_ada_layer_norm: lowercase = self.norma(A__ ,A__) elif self.use_ada_layer_norm_zero: lowercase , lowercase , lowercase , lowercase , lowercase = self.norma( A__ ,A__ ,A__ ,hidden_dtype=hidden_states.dtype) else: lowercase = self.norma(A__) lowercase = cross_attention_kwargs if cross_attention_kwargs is not None else {} lowercase = self.attna( A__ ,encoder_hidden_states=encoder_hidden_states if self.only_cross_attention else None ,attention_mask=A__ ,**A__ ,) if self.use_ada_layer_norm_zero: lowercase = gate_msa.unsqueeze(1) * attn_output lowercase = attn_output + hidden_states # 2. Cross-Attention if self.attna is not None: lowercase = ( self.norma(A__ ,A__) if self.use_ada_layer_norm else self.norma(A__) ) lowercase = self.attna( A__ ,encoder_hidden_states=A__ ,attention_mask=A__ ,**A__ ,) lowercase = attn_output + hidden_states # 3. Feed-forward lowercase = self.norma(A__) if self.use_ada_layer_norm_zero: lowercase = norm_hidden_states * (1 + scale_mlp[:, None]) + shift_mlp[:, None] if self._chunk_size is not None: # "feed_forward_chunk_size" can be used to save memory if norm_hidden_states.shape[self._chunk_dim] % self._chunk_size != 0: raise ValueError( f'`hidden_states` dimension to be chunked: {norm_hidden_states.shape[self._chunk_dim]} has to be divisible by chunk size: {self._chunk_size}. Make sure to set an appropriate `chunk_size` when calling `unet.enable_forward_chunking`.') lowercase = norm_hidden_states.shape[self._chunk_dim] // self._chunk_size lowercase = torch.cat( [self.ff(A__) for hid_slice in norm_hidden_states.chunk(A__ ,dim=self._chunk_dim)] ,dim=self._chunk_dim ,) else: lowercase = self.ff(A__) if self.use_ada_layer_norm_zero: lowercase = gate_mlp.unsqueeze(1) * ff_output lowercase = ff_output + hidden_states return hidden_states class lowercase ( nn.Module ): def __init__( self ,A__ ,A__ = None ,A__ = 4 ,A__ = 0.0 ,A__ = "geglu" ,A__ = False ,): super().__init__() lowercase = int(dim * mult) lowercase = dim_out if dim_out is not None else dim if activation_fn == "gelu": lowercase = GELU(A__ ,A__) if activation_fn == "gelu-approximate": lowercase = GELU(A__ ,A__ ,approximate='''tanh''') elif activation_fn == "geglu": lowercase = GEGLU(A__ ,A__) elif activation_fn == "geglu-approximate": lowercase = ApproximateGELU(A__ ,A__) lowercase = nn.ModuleList([]) # project in self.net.append(A__) # project dropout self.net.append(nn.Dropout(A__)) # project out self.net.append(nn.Linear(A__ ,A__)) # FF as used in Vision Transformer, MLP-Mixer, etc. have a final dropout if final_dropout: self.net.append(nn.Dropout(A__)) def A__ ( self ,A__): for module in self.net: lowercase = module(A__) return hidden_states class lowercase ( nn.Module ): def __init__( self ,A__ ,A__ ,A__ = "none"): super().__init__() lowercase = nn.Linear(A__ ,A__) lowercase = approximate def A__ ( self ,A__): if gate.device.type != "mps": return F.gelu(A__ ,approximate=self.approximate) # mps: gelu is not implemented for float16 return F.gelu(gate.to(dtype=torch.floataa) ,approximate=self.approximate).to(dtype=gate.dtype) def A__ ( self ,A__): lowercase = self.proj(A__) lowercase = self.gelu(A__) return hidden_states class lowercase ( nn.Module ): def __init__( self ,A__ ,A__): super().__init__() lowercase = nn.Linear(A__ ,dim_out * 2) def A__ ( self ,A__): if gate.device.type != "mps": return F.gelu(A__) # mps: gelu is not implemented for float16 return F.gelu(gate.to(dtype=torch.floataa)).to(dtype=gate.dtype) def A__ ( self ,A__): lowercase , lowercase = self.proj(A__).chunk(2 ,dim=-1) return hidden_states * self.gelu(A__) class lowercase ( nn.Module ): def __init__( self ,A__ ,A__): super().__init__() lowercase = nn.Linear(A__ ,A__) def A__ ( self ,A__): lowercase = self.proj(A__) return x * torch.sigmoid(1.702 * x) class lowercase ( nn.Module ): def __init__( self ,A__ ,A__): super().__init__() lowercase = nn.Embedding(A__ ,A__) lowercase = nn.SiLU() lowercase = nn.Linear(A__ ,embedding_dim * 2) lowercase = nn.LayerNorm(A__ ,elementwise_affine=A__) def A__ ( self ,A__ ,A__): lowercase = self.linear(self.silu(self.emb(A__))) lowercase , lowercase = torch.chunk(A__ ,2) lowercase = self.norm(A__) * (1 + scale) + shift return x class lowercase ( nn.Module ): def __init__( self ,A__ ,A__): super().__init__() lowercase = CombinedTimestepLabelEmbeddings(A__ ,A__) lowercase = nn.SiLU() lowercase = nn.Linear(A__ ,6 * embedding_dim ,bias=A__) lowercase = nn.LayerNorm(A__ ,elementwise_affine=A__ ,eps=1E-6) def A__ ( self ,A__ ,A__ ,A__ ,A__=None): lowercase = self.linear(self.silu(self.emb(A__ ,A__ ,hidden_dtype=A__))) lowercase , lowercase , lowercase , lowercase , lowercase , lowercase = emb.chunk(6 ,dim=1) lowercase = self.norm(A__) * (1 + scale_msa[:, None]) + shift_msa[:, None] return x, gate_msa, shift_mlp, scale_mlp, gate_mlp class lowercase ( nn.Module ): def __init__( self ,A__ ,A__ ,A__ ,A__ = None ,A__ = 1E-5): super().__init__() lowercase = num_groups lowercase = eps if act_fn is None: lowercase = None else: lowercase = get_activation(A__) lowercase = nn.Linear(A__ ,out_dim * 2) def A__ ( self ,A__ ,A__): if self.act: lowercase = self.act(A__) lowercase = self.linear(A__) lowercase = emb[:, :, None, None] lowercase , lowercase = emb.chunk(2 ,dim=1) lowercase = F.group_norm(A__ ,self.num_groups ,eps=self.eps) lowercase = x * (1 + scale) + shift return x
633
import warnings from diffusers import StableDiffusionInpaintPipeline as StableDiffusionInpaintPipeline # noqa F401 warnings.warn( "The `inpainting.py` script is outdated. Please use directly `from diffusers import" " StableDiffusionInpaintPipeline` instead." )
633
1
from multiprocessing import Lock, Pipe, Process # lock used to ensure that two processes do not access a pipe at the same time lowercase__ :List[str] = Lock() def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' global process_lock # we perform n swaps since after n swaps we know we are sorted # we *could* stop early if we are sorted already, but it takes as long to # find out we are sorted as it does to sort the list with this algorithm for i in range(0 , 10 ): if (i + position) % 2 == 0 and r_send is not None: # send your value to your right neighbor process_lock.acquire() r_send[1].send(lowerCAmelCase__ ) process_lock.release() # receive your right neighbor's value process_lock.acquire() lowercase = rr_cv[0].recv() process_lock.release() # take the lower value since you are on the left lowercase = min(lowerCAmelCase__ , lowerCAmelCase__ ) elif (i + position) % 2 != 0 and l_send is not None: # send your value to your left neighbor process_lock.acquire() l_send[1].send(lowerCAmelCase__ ) process_lock.release() # receive your left neighbor's value process_lock.acquire() lowercase = lr_cv[0].recv() process_lock.release() # take the higher value since you are on the right lowercase = max(lowerCAmelCase__ , lowerCAmelCase__ ) # after all swaps are performed, send the values back to main result_pipe[1].send(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = [] lowercase = [] # initialize the list of pipes where the values will be retrieved for _ in arr: result_pipe.append(Pipe() ) # creates the processes # the first and last process only have one neighbor so they are made outside # of the loop lowercase = Pipe() lowercase = Pipe() process_array_.append( Process( target=lowerCAmelCase__ , args=(0, arr[0], None, temp_rs, None, temp_rr, result_pipe[0]) , ) ) lowercase = temp_rs lowercase = temp_rr for i in range(1 , len(lowerCAmelCase__ ) - 1 ): lowercase = Pipe() lowercase = Pipe() process_array_.append( Process( target=lowerCAmelCase__ , args=(i, arr[i], temp_ls, temp_rs, temp_lr, temp_rr, result_pipe[i]) , ) ) lowercase = temp_rs lowercase = temp_rr process_array_.append( Process( target=lowerCAmelCase__ , args=( len(lowerCAmelCase__ ) - 1, arr[len(lowerCAmelCase__ ) - 1], temp_ls, None, temp_lr, None, result_pipe[len(lowerCAmelCase__ ) - 1], ) , ) ) # start the processes for p in process_array_: p.start() # wait for the processes to end and write their values to the list for p in range(0 , len(lowerCAmelCase__ ) ): lowercase = result_pipe[p][0].recv() process_array_[p].join() return arr def UpperCamelCase ( ): '''simple docstring''' lowercase = list(range(10 , 0 , -1 ) ) print('''Initial List''' ) print(*lowerCAmelCase__ ) lowercase = odd_even_transposition(lowerCAmelCase__ ) print('''Sorted List\n''' ) print(*lowerCAmelCase__ ) if __name__ == "__main__": main()
633
import random import unittest import numpy as np import transformers from transformers import is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax if is_flax_available(): import os import jax.numpy as jnp from jax import jit from transformers import AutoTokenizer, FlaxAutoModelForCausalLM from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model lowercase__ :Optional[Any] = "0.12" # assumed parallelism: 8 if is_torch_available(): import torch def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None ): '''simple docstring''' if rng is None: lowercase = random.Random() lowercase = 1 for dim in shape: total_dims *= dim lowercase = [] for _ in range(lowerCAmelCase__ ): values.append(rng.randint(0 , vocab_size - 1 ) ) lowercase = np.array(lowerCAmelCase__ , dtype=jnp.intaa ).reshape(lowerCAmelCase__ ) return output def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=None ): '''simple docstring''' lowercase = ids_tensor(lowerCAmelCase__ , vocab_size=2 , rng=lowerCAmelCase__ ) # make sure that at least one token is attended to for each batch lowercase = 1 return attn_mask @require_flax class lowercase : lowercase_ : Any =None lowercase_ : List[str] =() def A__ ( self): lowercase , lowercase = self.model_tester.prepare_config_and_inputs_for_common() # cut to half length & take max batch_size 3 lowercase = 2 lowercase = inputs['''input_ids'''].shape[-1] // 2 lowercase = inputs['''input_ids'''][:max_batch_size, :sequence_length] lowercase = jnp.ones_like(A__) lowercase = attention_mask[:max_batch_size, :sequence_length] # generate max 5 tokens lowercase = input_ids.shape[-1] + 5 if config.eos_token_id is not None and config.pad_token_id is None: # hack to allow generate for models such as GPT2 as is done in `generate()` lowercase = config.eos_token_id return config, input_ids, attention_mask, max_length @is_pt_flax_cross_test def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = False lowercase = max_length lowercase = 0 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model_class.__name__[4:] # Skip the "Flax" at the beginning lowercase = getattr(A__ ,A__) lowercase = pt_model_class(A__).eval() lowercase = load_flax_weights_in_pytorch_model(A__ ,flax_model.params) lowercase = flax_model.generate(A__).sequences lowercase = pt_model.generate(torch.tensor(A__ ,dtype=torch.long)) if flax_generation_outputs.shape[-1] > pt_generation_outputs.shape[-1]: lowercase = flax_generation_outputs[:, : pt_generation_outputs.shape[-1]] self.assertListEqual(pt_generation_outputs.numpy().tolist() ,flax_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = False lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = True lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = False lowercase = max_length lowercase = 2 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = False lowercase = max_length lowercase = 2 lowercase = 2 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[0] ,input_ids.shape[0] * config.num_return_sequences) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = True lowercase = max_length lowercase = 0.8 lowercase = 1_0 lowercase = 0.3 lowercase = 1 lowercase = 8 lowercase = 9 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = max_length lowercase = 1 lowercase = 8 lowercase = 9 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = max_length lowercase = 2 lowercase = 1 lowercase = 8 lowercase = 9 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() # pad attention mask on the left lowercase = attention_mask.at[(0, 0)].set(0) lowercase = False lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__ ,attention_mask=A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__ ,attention_mask=A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() # pad attention mask on the left lowercase = attention_mask.at[(0, 0)].set(0) lowercase = True lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__ ,attention_mask=A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__ ,attention_mask=A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() # pad attention mask on the left lowercase = attention_mask.at[(0, 0)].set(0) lowercase = 2 lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__ ,attention_mask=A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__ ,attention_mask=A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) @require_flax class lowercase ( unittest.TestCase ): def A__ ( self): lowercase = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-bert''') lowercase = FlaxAutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-bert-flax-only''') lowercase = '''Hello world''' lowercase = tokenizer(A__ ,return_tensors='''np''').input_ids # typos are quickly detected (the correct argument is `do_sample`) with self.assertRaisesRegex(A__ ,'''do_samples'''): model.generate(A__ ,do_samples=A__) # arbitrary arguments that will not be used anywhere are also not accepted with self.assertRaisesRegex(A__ ,'''foo'''): lowercase = {'''foo''': '''bar'''} model.generate(A__ ,**A__)
633
1
import argparse import os import torch from transformers import ( XLNetConfig, XLNetForQuestionAnswering, XLNetForSequenceClassification, XLNetLMHeadModel, load_tf_weights_in_xlnet, ) from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging lowercase__ :List[Any] = { "cola": 2, "mnli": 3, "mrpc": 2, "sst-2": 2, "sts-b": 1, "qqp": 2, "qnli": 2, "rte": 2, "wnli": 2, } logging.set_verbosity_info() def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None ): '''simple docstring''' # Initialise PyTorch model lowercase = XLNetConfig.from_json_file(lowerCAmelCase__ ) lowercase = finetuning_task.lower() if finetuning_task is not None else '''''' if finetuning_task in GLUE_TASKS_NUM_LABELS: print(f'Building PyTorch XLNetForSequenceClassification model from configuration: {config}' ) lowercase = finetuning_task lowercase = GLUE_TASKS_NUM_LABELS[finetuning_task] lowercase = XLNetForSequenceClassification(lowerCAmelCase__ ) elif "squad" in finetuning_task: lowercase = finetuning_task lowercase = XLNetForQuestionAnswering(lowerCAmelCase__ ) else: lowercase = XLNetLMHeadModel(lowerCAmelCase__ ) # Load weights from tf checkpoint load_tf_weights_in_xlnet(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # Save pytorch-model lowercase = os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase = os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) print(f'Save PyTorch model to {os.path.abspath(lowerCAmelCase__ )}' ) torch.save(model.state_dict() , lowerCAmelCase__ ) print(f'Save configuration file to {os.path.abspath(lowerCAmelCase__ )}' ) with open(lowerCAmelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": lowercase__ :Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--xlnet_config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained XLNet model. \n" "This specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the folder to store the PyTorch model or dataset/vocab.", ) parser.add_argument( "--finetuning_task", default=None, type=str, help="Name of a task on which the XLNet TensorFlow model was fine-tuned", ) lowercase__ :Tuple = parser.parse_args() print(args) convert_xlnet_checkpoint_to_pytorch( args.tf_checkpoint_path, args.xlnet_config_file, args.pytorch_dump_folder_path, args.finetuning_task )
633
import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : List[str] =(UniPCMultistepScheduler,) lowercase_ : Tuple =(('''num_inference_steps''', 25),) def A__ ( self ,**A__): lowercase = { '''num_train_timesteps''': 1_0_0_0, '''beta_start''': 0.0001, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', '''solver_order''': 2, '''solver_type''': '''bh2''', } config.update(**A__) return config def A__ ( self ,A__=0 ,**A__): lowercase = dict(self.forward_default_kwargs) lowercase = kwargs.pop('''num_inference_steps''' ,A__) lowercase = self.dummy_sample lowercase = 0.1 * sample lowercase = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) scheduler.set_timesteps(A__) # copy over dummy past residuals lowercase = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A__) lowercase = scheduler_class.from_pretrained(A__) new_scheduler.set_timesteps(A__) # copy over dummy past residuals lowercase = dummy_past_residuals[: new_scheduler.config.solver_order] lowercase , lowercase = sample, sample for t in range(A__ ,time_step + scheduler.config.solver_order + 1): lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = new_scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1E-5, "Scheduler outputs are not identical" def A__ ( self ,A__=0 ,**A__): lowercase = dict(self.forward_default_kwargs) lowercase = kwargs.pop('''num_inference_steps''' ,A__) lowercase = self.dummy_sample lowercase = 0.1 * sample lowercase = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config() lowercase = scheduler_class(**A__) scheduler.set_timesteps(A__) # copy over dummy past residuals (must be after setting timesteps) lowercase = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A__) lowercase = scheduler_class.from_pretrained(A__) # copy over dummy past residuals new_scheduler.set_timesteps(A__) # copy over dummy past residual (must be after setting timesteps) lowercase = dummy_past_residuals[: new_scheduler.config.solver_order] lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = new_scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1E-5, "Scheduler outputs are not identical" def A__ ( self ,A__=None ,**A__): if scheduler is None: lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) lowercase = 1_0 lowercase = self.dummy_model() lowercase = self.dummy_sample_deter scheduler.set_timesteps(A__) for i, t in enumerate(scheduler.timesteps): lowercase = model(A__ ,A__) lowercase = scheduler.step(A__ ,A__ ,A__).prev_sample return sample def A__ ( self): lowercase = dict(self.forward_default_kwargs) lowercase = kwargs.pop('''num_inference_steps''' ,A__) for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config() lowercase = scheduler_class(**A__) lowercase = self.dummy_sample lowercase = 0.1 * sample if num_inference_steps is not None and hasattr(A__ ,'''set_timesteps'''): scheduler.set_timesteps(A__) elif num_inference_steps is not None and not hasattr(A__ ,'''set_timesteps'''): lowercase = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) lowercase = [residual + 0.2, residual + 0.15, residual + 0.10] lowercase = dummy_past_residuals[: scheduler.config.solver_order] lowercase = scheduler.timesteps[5] lowercase = scheduler.timesteps[6] lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample self.assertEqual(output_a.shape ,sample.shape) self.assertEqual(output_a.shape ,output_a.shape) def A__ ( self): # make sure that iterating over schedulers with same config names gives same results # for defaults lowercase = UniPCMultistepScheduler(**self.get_scheduler_config()) lowercase = self.full_loop(scheduler=A__) lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.2464) < 1E-3 lowercase = DPMSolverSinglestepScheduler.from_config(scheduler.config) lowercase = DEISMultistepScheduler.from_config(scheduler.config) lowercase = DPMSolverMultistepScheduler.from_config(scheduler.config) lowercase = UniPCMultistepScheduler.from_config(scheduler.config) lowercase = self.full_loop(scheduler=A__) lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.2464) < 1E-3 def A__ ( self): for timesteps in [2_5, 5_0, 1_0_0, 9_9_9, 1_0_0_0]: self.check_over_configs(num_train_timesteps=A__) def A__ ( self): self.check_over_configs(thresholding=A__) for order in [1, 2, 3]: for solver_type in ["bh1", "bh2"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=A__ ,prediction_type=A__ ,sample_max_value=A__ ,solver_order=A__ ,solver_type=A__ ,) def A__ ( self): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=A__) def A__ ( self): for solver_type in ["bh1", "bh2"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=A__ ,solver_type=A__ ,prediction_type=A__ ,) lowercase = self.full_loop( solver_order=A__ ,solver_type=A__ ,prediction_type=A__ ,) assert not torch.isnan(A__).any(), "Samples have nan numbers" def A__ ( self): self.check_over_configs(lower_order_final=A__) self.check_over_configs(lower_order_final=A__) def A__ ( self): for num_inference_steps in [1, 2, 3, 5, 1_0, 5_0, 1_0_0, 9_9_9, 1_0_0_0]: self.check_over_forward(num_inference_steps=A__ ,time_step=0) def A__ ( self): lowercase = self.full_loop() lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.2464) < 1E-3 def A__ ( self): lowercase = self.full_loop(prediction_type='''v_prediction''') lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.1014) < 1E-3 def A__ ( self): lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config(thresholding=A__ ,dynamic_thresholding_ratio=0) lowercase = scheduler_class(**A__) lowercase = 1_0 lowercase = self.dummy_model() lowercase = self.dummy_sample_deter.half() scheduler.set_timesteps(A__) for i, t in enumerate(scheduler.timesteps): lowercase = model(A__ ,A__) lowercase = scheduler.step(A__ ,A__ ,A__).prev_sample assert sample.dtype == torch.floataa def A__ ( self ,**A__): for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) scheduler.set_timesteps(scheduler.config.num_train_timesteps) assert len(scheduler.timesteps.unique()) == scheduler.num_inference_steps
633
1
import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin from .feature_extraction_wavaveca import WavaVecaFeatureExtractor from .tokenization_wavaveca import WavaVecaCTCTokenizer class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : List[str] ='''Wav2Vec2FeatureExtractor''' lowercase_ : Optional[int] ='''AutoTokenizer''' def __init__( self ,A__ ,A__): super().__init__(A__ ,A__) lowercase = self.feature_extractor lowercase = False @classmethod def A__ ( cls ,A__ ,**A__): try: return super().from_pretrained(A__ ,**A__) except OSError: warnings.warn( f'Loading a tokenizer inside {cls.__name__} from a config that does not' ''' include a `tokenizer_class` attribute is deprecated and will be ''' '''removed in v5. Please add `\'tokenizer_class\': \'Wav2Vec2CTCTokenizer\'`''' ''' attribute to either your `config.json` or `tokenizer_config.json` ''' '''file to suppress this warning: ''' ,A__ ,) lowercase = WavaVecaFeatureExtractor.from_pretrained(A__ ,**A__) lowercase = WavaVecaCTCTokenizer.from_pretrained(A__ ,**A__) return cls(feature_extractor=A__ ,tokenizer=A__) def __call__( self ,*A__ ,**A__): # For backward compatibility if self._in_target_context_manager: return self.current_processor(*A__ ,**A__) if "raw_speech" in kwargs: warnings.warn('''Using `raw_speech` as a keyword argument is deprecated. Use `audio` instead.''') lowercase = kwargs.pop('''raw_speech''') else: lowercase = kwargs.pop('''audio''' ,A__) lowercase = kwargs.pop('''sampling_rate''' ,A__) lowercase = kwargs.pop('''text''' ,A__) if len(A__) > 0: lowercase = args[0] lowercase = args[1:] if audio is None and text is None: raise ValueError('''You need to specify either an `audio` or `text` input to process.''') if audio is not None: lowercase = self.feature_extractor(A__ ,*A__ ,sampling_rate=A__ ,**A__) if text is not None: lowercase = self.tokenizer(A__ ,**A__) if text is None: return inputs elif audio is None: return encodings else: lowercase = encodings['''input_ids'''] return inputs def A__ ( self ,*A__ ,**A__): # For backward compatibility if self._in_target_context_manager: return self.current_processor.pad(*A__ ,**A__) lowercase = kwargs.pop('''input_features''' ,A__) lowercase = kwargs.pop('''labels''' ,A__) if len(A__) > 0: lowercase = args[0] lowercase = args[1:] if input_features is not None: lowercase = self.feature_extractor.pad(A__ ,*A__ ,**A__) if labels is not None: lowercase = self.tokenizer.pad(A__ ,**A__) if labels is None: return input_features elif input_features is None: return labels else: lowercase = labels['''input_ids'''] return input_features def A__ ( self ,*A__ ,**A__): return self.tokenizer.batch_decode(*A__ ,**A__) def A__ ( self ,*A__ ,**A__): return self.tokenizer.decode(*A__ ,**A__) @contextmanager def A__ ( self): warnings.warn( '''`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ''' '''labels by using the argument `text` of the regular `__call__` method (either in the same call as ''' '''your audio inputs, or in a separate call.''') lowercase = True lowercase = self.tokenizer yield lowercase = self.feature_extractor lowercase = False
633
import argparse from collections import defaultdict import yaml lowercase__ :Optional[int] = "docs/source/en/_toctree.yml" def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = defaultdict(lowerCAmelCase__ ) for doc in model_doc: counts[doc["local"]] += 1 lowercase = [key for key, value in counts.items() if value > 1] lowercase = [] for duplicate_key in duplicates: lowercase = list({doc['''title'''] for doc in model_doc if doc['''local'''] == duplicate_key} ) if len(lowerCAmelCase__ ) > 1: raise ValueError( f'{duplicate_key} is present several times in the documentation table of content at ' '''`docs/source/en/_toctree.yml` with different *Title* values. Choose one of those and remove the ''' '''others.''' ) # Only add this once new_doc.append({'''local''': duplicate_key, '''title''': titles[0]} ) # Add none duplicate-keys new_doc.extend([doc for doc in model_doc if counts[doc['''local''']] == 1] ) # Sort return sorted(lowerCAmelCase__ , key=lambda lowerCAmelCase__ : s["title"].lower() ) def UpperCamelCase ( lowerCAmelCase__=False ): '''simple docstring''' with open(lowerCAmelCase__ , encoding='''utf-8''' ) as f: lowercase = yaml.safe_load(f.read() ) # Get to the API doc lowercase = 0 while content[api_idx]["title"] != "API": api_idx += 1 lowercase = content[api_idx]['''sections'''] # Then to the model doc lowercase = 0 while api_doc[model_idx]["title"] != "Models": model_idx += 1 lowercase = api_doc[model_idx]['''sections'''] lowercase = [(idx, section) for idx, section in enumerate(lowerCAmelCase__ ) if '''sections''' in section] lowercase = False for idx, modality_doc in modalities_docs: lowercase = modality_doc['''sections'''] lowercase = clean_model_doc_toc(lowerCAmelCase__ ) if old_modality_doc != new_modality_doc: lowercase = True if overwrite: lowercase = new_modality_doc if diff: if overwrite: lowercase = model_doc lowercase = api_doc with open(lowerCAmelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(yaml.dump(lowerCAmelCase__ , allow_unicode=lowerCAmelCase__ ) ) else: raise ValueError( '''The model doc part of the table of content is not properly sorted, run `make style` to fix this.''' ) if __name__ == "__main__": lowercase__ :Any = argparse.ArgumentParser() parser.add_argument("--fix_and_overwrite", action="store_true", help="Whether to fix inconsistencies.") lowercase__ :int = parser.parse_args() check_model_doc(args.fix_and_overwrite)
633
1
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BertTokenizer, BlipImageProcessor, BlipProcessor, PreTrainedTokenizerFast @require_vision class lowercase ( unittest.TestCase ): def A__ ( self): lowercase = tempfile.mkdtemp() lowercase = BlipImageProcessor() lowercase = BertTokenizer.from_pretrained('''hf-internal-testing/tiny-random-BertModel''') lowercase = BlipProcessor(A__ ,A__) processor.save_pretrained(self.tmpdirname) def A__ ( self ,**A__): return AutoProcessor.from_pretrained(self.tmpdirname ,**A__).tokenizer def A__ ( self ,**A__): return AutoProcessor.from_pretrained(self.tmpdirname ,**A__).image_processor def A__ ( self): shutil.rmtree(self.tmpdirname) def A__ ( self): lowercase = [np.random.randint(2_5_5 ,size=(3, 3_0, 4_0_0) ,dtype=np.uinta)] lowercase = [Image.fromarray(np.moveaxis(A__ ,0 ,-1)) for x in image_inputs] return image_inputs def A__ ( self): lowercase = BlipProcessor(tokenizer=self.get_tokenizer() ,image_processor=self.get_image_processor()) processor.save_pretrained(self.tmpdirname) lowercase = self.get_tokenizer(bos_token='''(BOS)''' ,eos_token='''(EOS)''') lowercase = self.get_image_processor(do_normalize=A__ ,padding_value=1.0) lowercase = BlipProcessor.from_pretrained( self.tmpdirname ,bos_token='''(BOS)''' ,eos_token='''(EOS)''' ,do_normalize=A__ ,padding_value=1.0) self.assertEqual(processor.tokenizer.get_vocab() ,tokenizer_add_kwargs.get_vocab()) self.assertIsInstance(processor.tokenizer ,A__) self.assertEqual(processor.image_processor.to_json_string() ,image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor ,A__) def A__ ( self): lowercase = self.get_image_processor() lowercase = self.get_tokenizer() lowercase = BlipProcessor(tokenizer=A__ ,image_processor=A__) lowercase = self.prepare_image_inputs() lowercase = image_processor(A__ ,return_tensors='''np''') lowercase = processor(images=A__ ,return_tensors='''np''') for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1E-2) def A__ ( self): lowercase = self.get_image_processor() lowercase = self.get_tokenizer() lowercase = BlipProcessor(tokenizer=A__ ,image_processor=A__) lowercase = '''lower newer''' lowercase = processor(text=A__) lowercase = tokenizer(A__ ,return_token_type_ids=A__) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] ,encoded_processor[key]) def A__ ( self): lowercase = self.get_image_processor() lowercase = self.get_tokenizer() lowercase = BlipProcessor(tokenizer=A__ ,image_processor=A__) lowercase = '''lower newer''' lowercase = self.prepare_image_inputs() lowercase = processor(text=A__ ,images=A__) self.assertListEqual(list(inputs.keys()) ,['''pixel_values''', '''input_ids''', '''attention_mask''']) # test if it raises when no input is passed with pytest.raises(A__): processor() def A__ ( self): lowercase = self.get_image_processor() lowercase = self.get_tokenizer() lowercase = BlipProcessor(tokenizer=A__ ,image_processor=A__) lowercase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] lowercase = processor.batch_decode(A__) lowercase = tokenizer.batch_decode(A__) self.assertListEqual(A__ ,A__) def A__ ( self): lowercase = self.get_image_processor() lowercase = self.get_tokenizer() lowercase = BlipProcessor(tokenizer=A__ ,image_processor=A__) lowercase = '''lower newer''' lowercase = self.prepare_image_inputs() lowercase = processor(text=A__ ,images=A__) # For now the processor supports only ['pixel_values', 'input_ids', 'attention_mask'] self.assertListEqual(list(inputs.keys()) ,['''pixel_values''', '''input_ids''', '''attention_mask'''])
633
from __future__ import annotations from decimal import Decimal from math import * # noqa: F403 from sympy import diff def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 10**-10 ): '''simple docstring''' lowercase = a while True: lowercase = Decimal(lowerCAmelCase__ ) - ( Decimal(eval(lowerCAmelCase__ ) ) / Decimal(eval(str(diff(lowerCAmelCase__ ) ) ) ) # noqa: S307 ) # This number dictates the accuracy of the answer if abs(eval(lowerCAmelCase__ ) ) < precision: # noqa: S307 return float(lowerCAmelCase__ ) # Let's Execute if __name__ == "__main__": # Find root of trigonometric function # Find value of pi print(F'The root of sin(x) = 0 is {newton_raphson("sin(x)", 2)}') # Find root of polynomial print(F'The root of x**2 - 5*x + 2 = 0 is {newton_raphson("x**2 - 5*x + 2", 0.4)}') # Find Square Root of 5 print(F'The root of log(x) - 1 = 0 is {newton_raphson("log(x) - 1", 2)}') # Exponential Roots print(F'The root of exp(x) - 1 = 0 is {newton_raphson("exp(x) - 1", 0)}')
633
1
def UpperCamelCase ( lowerCAmelCase__ = 100_0000 ): '''simple docstring''' lowercase = limit + 1 lowercase = [0] * limit for first_term in range(1 , lowerCAmelCase__ ): for n in range(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): lowercase = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a lowercase = sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(F'{solution() = }')
633
from pathlib import Path import numpy as np from PIL import Image def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase , lowercase = rgb[:, :, 0], rgb[:, :, 1], rgb[:, :, 2] return 0.29_89 * r + 0.58_70 * g + 0.11_40 * b def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return (gray > 127) & (gray <= 255) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = np.zeros_like(lowerCAmelCase__ ) lowercase = np.zeros( (image.shape[0] + kernel.shape[0] - 1, image.shape[1] + kernel.shape[1] - 1) ) # Copy image to padded image lowercase = image # Iterate over image & apply kernel for x in range(image.shape[1] ): for y in range(image.shape[0] ): lowercase = ( kernel * image_padded[y : y + kernel.shape[0], x : x + kernel.shape[1]] ).sum() lowercase = int(summation > 0 ) return output if __name__ == "__main__": # read original image lowercase__ :str = Path(__file__).resolve().parent / "image_data" / "lena.jpg" lowercase__ :List[str] = np.array(Image.open(lena_path)) # kernel to be applied lowercase__ :Union[str, Any] = np.array([[0, 1, 0], [1, 1, 1], [0, 1, 0]]) lowercase__ :Optional[int] = dilation(gray_to_binary(rgb_to_gray(lena)), structuring_element) # Save the output image lowercase__ :str = Image.fromarray(output).convert("RGB") pil_img.save("result_dilation.png")
633
1
import os def UpperCamelCase ( ): '''simple docstring''' with open(os.path.dirname(lowerCAmelCase__ ) + '''/p022_names.txt''' ) as file: lowercase = str(file.readlines()[0] ) lowercase = names.replace('''"''' , '''''' ).split(''',''' ) names.sort() lowercase = 0 lowercase = 0 for i, name in enumerate(lowerCAmelCase__ ): for letter in name: name_score += ord(lowerCAmelCase__ ) - 64 total_score += (i + 1) * name_score lowercase = 0 return total_score if __name__ == "__main__": print(solution())
633
import argparse import torch from torch import nn from transformers import MaMaaaConfig, MaMaaaForConditionalGeneration def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = [ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''decoder.output_projection.weight''', '''_float_tensor''', '''encoder.embed_positions._float_tensor''', '''decoder.embed_positions._float_tensor''', ] for k in ignore_keys: state_dict.pop(lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase = emb.weight.shape lowercase = nn.Linear(lowerCAmelCase__ , lowerCAmelCase__ , bias=lowerCAmelCase__ ) lowercase = emb.weight.data return lin_layer def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = torch.load(lowerCAmelCase__ , map_location='''cpu''' ) lowercase = mam_aaa['''args'''] or mam_aaa['''cfg''']['''model'''] lowercase = mam_aaa['''model'''] remove_ignore_keys_(lowerCAmelCase__ ) lowercase = state_dict['''encoder.embed_tokens.weight'''].shape[0] lowercase = MaMaaaConfig( vocab_size=lowerCAmelCase__ , max_position_embeddings=1024 , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , encoder_layerdrop=args.encoder_layerdrop , decoder_layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='''relu''' , ) lowercase = state_dict['''decoder.embed_tokens.weight'''] lowercase = MaMaaaForConditionalGeneration(lowerCAmelCase__ ) model.model.load_state_dict(lowerCAmelCase__ , strict=lowerCAmelCase__ ) lowercase = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": lowercase__ :Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument("fairseq_path", type=str, help="path to a model.pt on local filesystem.") parser.add_argument("pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") lowercase__ :Tuple = parser.parse_args() lowercase__ :int = convert_fairseq_mamaaa_checkpoint_from_disk(args.fairseq_pathß) model.save_pretrained(args.pytorch_dump_folder_path)
633
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) lowercase__ :Any = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :int = ["NllbTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :str = ["NllbTokenizerFast"] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb import NllbTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb_fast import NllbTokenizerFast else: import sys lowercase__ :Union[str, Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
633
from __future__ import annotations from random import random class lowercase : def __init__( self ,A__ = None): lowercase = value lowercase = random() lowercase = None lowercase = None def __repr__( self): from pprint import pformat if self.left is None and self.right is None: return f'\'{self.value}: {self.prior:.5}\'' else: return pformat( {f'{self.value}: {self.prior:.5}': (self.left, self.right)} ,indent=1) def __str__( self): lowercase = str(self.value) + ''' ''' lowercase = str(self.left or '''''') lowercase = str(self.right or '''''') return value + left + right def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' if root is None: # None tree is split into 2 Nones return None, None elif root.value is None: return None, None else: if value < root.value: lowercase , lowercase = split(root.left , lowerCAmelCase__ ) return left, root else: lowercase , lowercase = split(root.right , lowerCAmelCase__ ) return root, right def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' if (not left) or (not right): # If one node is None, return the other return left or right elif left.prior < right.prior: lowercase = merge(left.right , lowerCAmelCase__ ) return left else: lowercase = merge(lowerCAmelCase__ , right.left ) return right def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = Node(lowerCAmelCase__ ) lowercase , lowercase = split(lowerCAmelCase__ , lowerCAmelCase__ ) return merge(merge(lowerCAmelCase__ , lowerCAmelCase__ ) , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase = split(lowerCAmelCase__ , value - 1 ) lowercase , lowercase = split(lowerCAmelCase__ , lowerCAmelCase__ ) return merge(lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if not root: # None return else: inorder(root.left ) print(root.value , end=''',''' ) inorder(root.right ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' for arg in args.split(): if arg[0] == "+": lowercase = insert(lowerCAmelCase__ , int(arg[1:] ) ) elif arg[0] == "-": lowercase = erase(lowerCAmelCase__ , int(arg[1:] ) ) else: print('''Unknown command''' ) return root def UpperCamelCase ( ): '''simple docstring''' lowercase = None print( '''enter numbers to create a tree, + value to add value into treap, ''' '''- value to erase all nodes with value. \'q\' to quit. ''' ) lowercase = input() while args != "q": lowercase = interact_treap(lowerCAmelCase__ , lowerCAmelCase__ ) print(lowerCAmelCase__ ) lowercase = input() print('''good by!''' ) if __name__ == "__main__": import doctest doctest.testmod() main()
633
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowercase__ :Tuple = {"configuration_yolos": ["YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP", "YolosConfig", "YolosOnnxConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :Tuple = ["YolosFeatureExtractor"] lowercase__ :Optional[int] = ["YolosImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :Optional[int] = [ "YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST", "YolosForObjectDetection", "YolosModel", "YolosPreTrainedModel", ] if TYPE_CHECKING: from .configuration_yolos import YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP, YolosConfig, YolosOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_yolos import YolosFeatureExtractor from .image_processing_yolos import YolosImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_yolos import ( YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST, YolosForObjectDetection, YolosModel, YolosPreTrainedModel, ) else: import sys lowercase__ :int = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
633
def UpperCamelCase ( lowerCAmelCase__ = 1000 ): '''simple docstring''' lowercase = -1 lowercase = 0 for a in range(1 , n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N eliminating c lowercase = (n * n - 2 * a * n) // (2 * n - 2 * a) lowercase = n - a - b if c * c == (a * a + b * b): lowercase = a * b * c if candidate >= product: lowercase = candidate return product if __name__ == "__main__": print(F'{solution() = }')
633
1
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision import transforms from transformers import BitImageProcessor, FocalNetConfig, FocalNetForImageClassification from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = [2, 2, 6, 2] if '''tiny''' in model_name else [2, 2, 18, 2] lowercase = True if '''large''' in model_name or '''huge''' in model_name else False lowercase = True if '''large''' in model_name or '''huge''' in model_name else False lowercase = True if '''large''' in model_name or '''huge''' in model_name else False if "large" in model_name or "xlarge" in model_name or "huge" in model_name: if "fl3" in model_name: lowercase = [3, 3, 3, 3] lowercase = [5, 5, 5, 5] elif "fl4" in model_name: lowercase = [4, 4, 4, 4] lowercase = [3, 3, 3, 3] if "tiny" in model_name or "small" in model_name or "base" in model_name: lowercase = [3, 3, 3, 3] if "lrf" in model_name: lowercase = [3, 3, 3, 3] else: lowercase = [2, 2, 2, 2] if "tiny" in model_name: lowercase = 96 elif "small" in model_name: lowercase = 96 elif "base" in model_name: lowercase = 128 elif "large" in model_name: lowercase = 192 elif "xlarge" in model_name: lowercase = 256 elif "huge" in model_name: lowercase = 352 # set label information lowercase = '''huggingface/label-files''' if "large" in model_name or "huge" in model_name: lowercase = '''imagenet-22k-id2label.json''' else: lowercase = '''imagenet-1k-id2label.json''' lowercase = json.load(open(hf_hub_download(lowerCAmelCase__ , lowerCAmelCase__ , repo_type='''dataset''' ) , '''r''' ) ) lowercase = {int(lowerCAmelCase__ ): v for k, v in idalabel.items()} lowercase = {v: k for k, v in idalabel.items()} lowercase = FocalNetConfig( embed_dim=lowerCAmelCase__ , depths=lowerCAmelCase__ , focal_levels=lowerCAmelCase__ , focal_windows=lowerCAmelCase__ , use_conv_embed=lowerCAmelCase__ , idalabel=lowerCAmelCase__ , labelaid=lowerCAmelCase__ , use_post_layernorm=lowerCAmelCase__ , use_layerscale=lowerCAmelCase__ , ) return config def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if "patch_embed.proj" in name: lowercase = name.replace('''patch_embed.proj''' , '''embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: lowercase = name.replace('''patch_embed.norm''' , '''embeddings.norm''' ) if "layers" in name: lowercase = '''encoder.''' + name if "encoder.layers" in name: lowercase = name.replace('''encoder.layers''' , '''encoder.stages''' ) if "downsample.proj" in name: lowercase = name.replace('''downsample.proj''' , '''downsample.projection''' ) if "blocks" in name: lowercase = name.replace('''blocks''' , '''layers''' ) if "modulation.f.weight" in name or "modulation.f.bias" in name: lowercase = name.replace('''modulation.f''' , '''modulation.projection_in''' ) if "modulation.h.weight" in name or "modulation.h.bias" in name: lowercase = name.replace('''modulation.h''' , '''modulation.projection_context''' ) if "modulation.proj.weight" in name or "modulation.proj.bias" in name: lowercase = name.replace('''modulation.proj''' , '''modulation.projection_out''' ) if name == "norm.weight": lowercase = '''layernorm.weight''' if name == "norm.bias": lowercase = '''layernorm.bias''' if "head" in name: lowercase = name.replace('''head''' , '''classifier''' ) else: lowercase = '''focalnet.''' + name return name def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=False ): '''simple docstring''' # fmt: off lowercase = { '''focalnet-tiny''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_srf.pth''', '''focalnet-tiny-lrf''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_lrf.pth''', '''focalnet-small''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_srf.pth''', '''focalnet-small-lrf''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_lrf.pth''', '''focalnet-base''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_srf.pth''', '''focalnet-base-lrf''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_lrf.pth''', '''focalnet-large-lrf-fl3''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384.pth''', '''focalnet-large-lrf-fl4''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384_fl4.pth''', '''focalnet-xlarge-lrf-fl3''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384.pth''', '''focalnet-xlarge-lrf-fl4''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384_fl4.pth''', } # fmt: on lowercase = model_name_to_url[model_name] print('''Checkpoint URL: ''' , lowerCAmelCase__ ) lowercase = torch.hub.load_state_dict_from_url(lowerCAmelCase__ , map_location='''cpu''' )['''model'''] # rename keys for key in state_dict.copy().keys(): lowercase = state_dict.pop(lowerCAmelCase__ ) lowercase = val lowercase = get_focalnet_config(lowerCAmelCase__ ) lowercase = FocalNetForImageClassification(lowerCAmelCase__ ) model.eval() # load state dict model.load_state_dict(lowerCAmelCase__ ) # verify conversion lowercase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' lowercase = BitImageProcessor( do_resize=lowerCAmelCase__ , size={'''shortest_edge''': 256} , resample=PILImageResampling.BILINEAR , do_center_crop=lowerCAmelCase__ , crop_size=224 , do_normalize=lowerCAmelCase__ , image_mean=lowerCAmelCase__ , image_std=lowerCAmelCase__ , ) lowercase = Image.open(requests.get(lowerCAmelCase__ , stream=lowerCAmelCase__ ).raw ) lowercase = processor(images=lowerCAmelCase__ , return_tensors='''pt''' ) lowercase = transforms.Compose( [ transforms.Resize(256 ), transforms.CenterCrop(224 ), transforms.ToTensor(), transforms.Normalize(mean=[0.4_85, 0.4_56, 0.4_06] , std=[0.2_29, 0.2_24, 0.2_25] ), ] ) lowercase = image_transforms(lowerCAmelCase__ ).unsqueeze(0 ) # verify pixel_values assert torch.allclose(inputs.pixel_values , lowerCAmelCase__ , atol=1E-4 ) lowercase = model(**lowerCAmelCase__ ) lowercase = outputs.logits.argmax(-1 ).item() print('''Predicted class:''' , model.config.idalabel[predicted_class_idx] ) print('''First values of logits:''' , outputs.logits[0, :3] ) if model_name == "focalnet-tiny": lowercase = torch.tensor([0.21_66, -0.43_68, 0.21_91] ) elif model_name == "focalnet-tiny-lrf": lowercase = torch.tensor([1.16_69, 0.01_25, -0.16_95] ) elif model_name == "focalnet-small": lowercase = torch.tensor([0.49_17, -0.04_30, 0.13_41] ) elif model_name == "focalnet-small-lrf": lowercase = torch.tensor([-0.25_88, -0.53_42, -0.23_31] ) elif model_name == "focalnet-base": lowercase = torch.tensor([-0.16_55, -0.40_90, -0.17_30] ) elif model_name == "focalnet-base-lrf": lowercase = torch.tensor([0.53_06, -0.04_83, -0.39_28] ) assert torch.allclose(outputs.logits[0, :3] , lowerCAmelCase__ , atol=1E-4 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(f'Saving model and processor of {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(lowerCAmelCase__ ) processor.save_pretrained(lowerCAmelCase__ ) if push_to_hub: print(f'Pushing model and processor of {model_name} to the hub...' ) model.push_to_hub(f'{model_name}' ) processor.push_to_hub(f'{model_name}' ) if __name__ == "__main__": lowercase__ :Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="focalnet-tiny", type=str, help="Name of the FocalNet model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to push the model and processor to the hub.", ) lowercase__ :Any = parser.parse_args() convert_focalnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
633
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ :Tuple = { "configuration_instructblip": [ "INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "InstructBlipConfig", "InstructBlipQFormerConfig", "InstructBlipVisionConfig", ], "processing_instructblip": ["InstructBlipProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :List[str] = [ "INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "InstructBlipQFormerModel", "InstructBlipPreTrainedModel", "InstructBlipForConditionalGeneration", "InstructBlipVisionModel", ] if TYPE_CHECKING: from .configuration_instructblip import ( INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, InstructBlipConfig, InstructBlipQFormerConfig, InstructBlipVisionConfig, ) from .processing_instructblip import InstructBlipProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_instructblip import ( INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST, InstructBlipForConditionalGeneration, InstructBlipPreTrainedModel, InstructBlipQFormerModel, InstructBlipVisionModel, ) else: import sys lowercase__ :List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
633
1
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import convert_to_rgb, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowercase__ :List[Any] = logging.get_logger(__name__) class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Union[str, Any] =['''pixel_values'''] def __init__( self ,A__ = True ,A__ = None ,A__ = PILImageResampling.BICUBIC ,A__ = True ,A__ = 1 / 2_5_5 ,A__ = True ,A__ = None ,A__ = None ,A__ = True ,**A__ ,): super().__init__(**A__) lowercase = size if size is not None else {'''height''': 3_8_4, '''width''': 3_8_4} lowercase = get_size_dict(A__ ,default_to_square=A__) lowercase = do_resize lowercase = size lowercase = resample lowercase = do_rescale lowercase = rescale_factor lowercase = do_normalize lowercase = image_mean if image_mean is not None else OPENAI_CLIP_MEAN lowercase = image_std if image_std is not None else OPENAI_CLIP_STD lowercase = do_convert_rgb def A__ ( self ,A__ ,A__ ,A__ = PILImageResampling.BICUBIC ,A__ = None ,**A__ ,): lowercase = get_size_dict(A__ ,default_to_square=A__) if "height" not in size or "width" not in size: raise ValueError(f'The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}') lowercase = (size['''height'''], size['''width''']) return resize(A__ ,size=A__ ,resample=A__ ,data_format=A__ ,**A__) def A__ ( self ,A__ ,A__ ,A__ = None ,**A__ ,): return rescale(A__ ,scale=A__ ,data_format=A__ ,**A__) def A__ ( self ,A__ ,A__ ,A__ ,A__ = None ,**A__ ,): return normalize(A__ ,mean=A__ ,std=A__ ,data_format=A__ ,**A__) def A__ ( self ,A__ ,A__ = None ,A__ = None ,A__ = None ,A__ = None ,A__ = None ,A__ = None ,A__ = None ,A__ = None ,A__ = None ,A__ = None ,A__ = ChannelDimension.FIRST ,**A__ ,): lowercase = do_resize if do_resize is not None else self.do_resize lowercase = resample if resample is not None else self.resample lowercase = do_rescale if do_rescale is not None else self.do_rescale lowercase = rescale_factor if rescale_factor is not None else self.rescale_factor lowercase = do_normalize if do_normalize is not None else self.do_normalize lowercase = image_mean if image_mean is not None else self.image_mean lowercase = image_std if image_std is not None else self.image_std lowercase = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb lowercase = size if size is not None else self.size lowercase = get_size_dict(A__ ,default_to_square=A__) lowercase = make_list_of_images(A__) if not valid_images(A__): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''') if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''') if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''') if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''') # PIL RGBA images are converted to RGB if do_convert_rgb: lowercase = [convert_to_rgb(A__) for image in images] # All transformations expect numpy arrays. lowercase = [to_numpy_array(A__) for image in images] if do_resize: lowercase = [self.resize(image=A__ ,size=A__ ,resample=A__) for image in images] if do_rescale: lowercase = [self.rescale(image=A__ ,scale=A__) for image in images] if do_normalize: lowercase = [self.normalize(image=A__ ,mean=A__ ,std=A__) for image in images] lowercase = [to_channel_dimension_format(A__ ,A__) for image in images] lowercase = BatchFeature(data={'''pixel_values''': images} ,tensor_type=A__) return encoded_outputs
633
from numpy import exp, pi, sqrt def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ = 0.0 , lowerCAmelCase__ = 1.0 ): '''simple docstring''' return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
633
1
def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = '''''' for ch in key: if ch == " " or ch not in key_no_dups and ch.isalpha(): key_no_dups += ch return key_no_dups def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = [chr(i + 65 ) for i in range(26 )] # Remove duplicate characters from key lowercase = remove_duplicates(key.upper() ) lowercase = len(lowerCAmelCase__ ) # First fill cipher with key characters lowercase = {alphabet[i]: char for i, char in enumerate(lowerCAmelCase__ )} # Then map remaining characters in alphabet to # the alphabet from the beginning for i in range(len(lowerCAmelCase__ ) , 26 ): lowercase = alphabet[i - offset] # Ensure we are not mapping letters to letters previously mapped while char in key: offset -= 1 lowercase = alphabet[i - offset] lowercase = char return cipher_alphabet def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' return "".join(cipher_map.get(lowerCAmelCase__ , lowerCAmelCase__ ) for ch in message.upper() ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = {v: k for k, v in cipher_map.items()} return "".join(rev_cipher_map.get(lowerCAmelCase__ , lowerCAmelCase__ ) for ch in message.upper() ) def UpperCamelCase ( ): '''simple docstring''' lowercase = input('''Enter message to encode or decode: ''' ).strip() lowercase = input('''Enter keyword: ''' ).strip() lowercase = input('''Encipher or decipher? E/D:''' ).strip()[0].lower() try: lowercase = {'''e''': encipher, '''d''': decipher}[option] except KeyError: raise KeyError('''invalid input option''' ) lowercase = create_cipher_map(lowerCAmelCase__ ) print(func(lowerCAmelCase__ , lowerCAmelCase__ ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
633
import asyncio import os import shutil import subprocess import sys import tempfile import unittest from distutils.util import strtobool from functools import partial from pathlib import Path from typing import List, Union from unittest import mock import torch from ..state import AcceleratorState, PartialState from ..utils import ( gather, is_bnb_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_mps_available, is_safetensors_available, is_tensorboard_available, is_torch_version, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=False ): '''simple docstring''' try: lowercase = os.environ[key] except KeyError: # KEY isn't set, default to `default`. lowercase = default else: # KEY is set, convert it to True or False. try: lowercase = strtobool(lowerCAmelCase__ ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(f'If set, {key} must be yes or no.' ) return _value lowercase__ :Dict = parse_flag_from_env("RUN_SLOW", default=False) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skip('''Test was skipped''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(_run_slow_tests , '''test is slow''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(not torch.cuda.is_available() , '''test requires only a CPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.cuda.is_available() , '''test requires a GPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_xpu_available() , '''test requires a XPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_mps_available() , '''test requires a `mps` backend support in `torch`''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless( is_transformers_available() and is_datasets_available() , '''test requires the Hugging Face suite''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_bnb_available() , '''test requires the bitsandbytes library''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_tpu_available() , '''test requires TPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.cuda.device_count() == 1 , '''test requires a GPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.xpu.device_count() == 1 , '''test requires a XPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.cuda.device_count() > 1 , '''test requires multiple GPUs''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.xpu.device_count() > 1 , '''test requires multiple XPUs''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_safetensors_available() , '''test requires safetensors''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_deepspeed_available() , '''test requires DeepSpeed''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_torch_version('''>=''' , '''1.12.0''' ) , '''test requires torch version >= 1.12.0''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__=None , lowerCAmelCase__=None ): '''simple docstring''' if test_case is None: return partial(lowerCAmelCase__ , version=lowerCAmelCase__ ) return unittest.skipUnless(is_torch_version('''>=''' , lowerCAmelCase__ ) , f'test requires torch version >= {version}' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_tensorboard_available() , '''test requires Tensorboard''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_wandb_available() , '''test requires wandb''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_comet_ml_available() , '''test requires comet_ml''' )(lowerCAmelCase__ ) lowercase__ :Dict = ( any([is_wandb_available(), is_tensorboard_available()]) and not is_comet_ml_available() ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless( _atleast_one_tracker_available , '''test requires at least one tracker to be available and for `comet_ml` to not be installed''' , )(lowerCAmelCase__ ) class lowercase ( unittest.TestCase ): lowercase_ : int =True @classmethod def A__ ( cls): lowercase = tempfile.mkdtemp() @classmethod def A__ ( cls): if os.path.exists(cls.tmpdir): shutil.rmtree(cls.tmpdir) def A__ ( self): if self.clear_on_setup: for path in Path(self.tmpdir).glob('''**/*'''): if path.is_file(): path.unlink() elif path.is_dir(): shutil.rmtree(A__) class lowercase ( unittest.TestCase ): def A__ ( self): super().tearDown() # Reset the state of the AcceleratorState singleton. AcceleratorState._reset_state() PartialState._reset_state() class lowercase ( unittest.TestCase ): def A__ ( self ,A__): lowercase = mocks if isinstance(A__ ,(tuple, list)) else [mocks] for m in self.mocks: m.start() self.addCleanup(m.stop) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = AcceleratorState() lowercase = tensor[None].clone().to(state.device ) lowercase = gather(lowerCAmelCase__ ).cpu() lowercase = tensor[0].cpu() for i in range(tensors.shape[0] ): if not torch.equal(tensors[i] , lowerCAmelCase__ ): return False return True class lowercase : def __init__( self ,A__ ,A__ ,A__): lowercase = returncode lowercase = stdout lowercase = stderr async def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' while True: lowercase = await stream.readline() if line: callback(lowerCAmelCase__ ) else: break async def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=False , lowerCAmelCase__=False ): '''simple docstring''' if echo: print('''\nRunning: ''' , ''' '''.join(lowerCAmelCase__ ) ) lowercase = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=lowerCAmelCase__ , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=lowerCAmelCase__ , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) lowercase = [] lowercase = [] def tee(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__="" ): lowercase = line.decode('''utf-8''' ).rstrip() sink.append(lowerCAmelCase__ ) if not quiet: print(lowerCAmelCase__ , lowerCAmelCase__ , file=lowerCAmelCase__ ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ asyncio.create_task(_read_stream(p.stdout , lambda lowerCAmelCase__ : tee(lowerCAmelCase__ , lowerCAmelCase__ , sys.stdout , label='''stdout:''' ) ) ), asyncio.create_task(_read_stream(p.stderr , lambda lowerCAmelCase__ : tee(lowerCAmelCase__ , lowerCAmelCase__ , sys.stderr , label='''stderr:''' ) ) ), ] , timeout=lowerCAmelCase__ , ) return _RunOutput(await p.wait() , lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=180 , lowerCAmelCase__=False , lowerCAmelCase__=True ): '''simple docstring''' lowercase = asyncio.get_event_loop() lowercase = loop.run_until_complete( _stream_subprocess(lowerCAmelCase__ , env=lowerCAmelCase__ , stdin=lowerCAmelCase__ , timeout=lowerCAmelCase__ , quiet=lowerCAmelCase__ , echo=lowerCAmelCase__ ) ) lowercase = ''' '''.join(lowerCAmelCase__ ) if result.returncode > 0: lowercase = '''\n'''.join(result.stderr ) raise RuntimeError( f'\'{cmd_str}\' failed with returncode {result.returncode}\n\n' f'The combined stderr from workers follows:\n{stderr}' ) return result class lowercase ( SCREAMING_SNAKE_CASE__ ): pass def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=False ): '''simple docstring''' try: lowercase = subprocess.check_output(lowerCAmelCase__ , stderr=subprocess.STDOUT ) if return_stdout: if hasattr(lowerCAmelCase__ , '''decode''' ): lowercase = output.decode('''utf-8''' ) return output except subprocess.CalledProcessError as e: raise SubprocessCallException( f'Command `{" ".join(lowerCAmelCase__ )}` failed with the following error:\n\n{e.output.decode()}' ) from e
633
1
import unittest from transformers import AutoTokenizer, NystromformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, NystromformerModel, ) from transformers.models.nystromformer.modeling_nystromformer import NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST class lowercase : def __init__( self ,A__ ,A__=1_3 ,A__=7 ,A__=True ,A__=True ,A__=True ,A__=True ,A__=9_9 ,A__=3_2 ,A__=5 ,A__=4 ,A__=3_7 ,A__="gelu" ,A__=0.1 ,A__=0.1 ,A__=5_1_2 ,A__=1_6 ,A__=2 ,A__=0.02 ,A__=3 ,A__=4 ,A__=None ,): lowercase = parent lowercase = batch_size lowercase = seq_length lowercase = is_training lowercase = use_input_mask lowercase = use_token_type_ids lowercase = use_labels lowercase = vocab_size lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_attention_heads lowercase = intermediate_size lowercase = hidden_act lowercase = hidden_dropout_prob lowercase = attention_probs_dropout_prob lowercase = max_position_embeddings lowercase = type_vocab_size lowercase = type_sequence_label_size lowercase = initializer_range lowercase = num_labels lowercase = num_choices lowercase = scope def A__ ( self): lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size) lowercase = None if self.use_input_mask: lowercase = random_attention_mask([self.batch_size, self.seq_length]) lowercase = None if self.use_token_type_ids: lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size) lowercase = None lowercase = None lowercase = None if self.use_labels: lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size) lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels) lowercase = ids_tensor([self.batch_size] ,self.num_choices) lowercase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A__ ( self): return NystromformerConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,is_decoder=A__ ,initializer_range=self.initializer_range ,) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__): lowercase = NystromformerModel(config=A__) model.to(A__) model.eval() lowercase = model(A__ ,attention_mask=A__ ,token_type_ids=A__) lowercase = model(A__ ,token_type_ids=A__) lowercase = model(A__) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size)) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__): lowercase = NystromformerForMaskedLM(config=A__) model.to(A__) model.eval() lowercase = model(A__ ,attention_mask=A__ ,token_type_ids=A__ ,labels=A__) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size)) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__): lowercase = NystromformerForQuestionAnswering(config=A__) model.to(A__) model.eval() lowercase = model( A__ ,attention_mask=A__ ,token_type_ids=A__ ,start_positions=A__ ,end_positions=A__ ,) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length)) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__): lowercase = self.num_labels lowercase = NystromformerForSequenceClassification(A__) model.to(A__) model.eval() lowercase = model(A__ ,attention_mask=A__ ,token_type_ids=A__ ,labels=A__) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels)) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__): lowercase = self.num_labels lowercase = NystromformerForTokenClassification(config=A__) model.to(A__) model.eval() lowercase = model(A__ ,attention_mask=A__ ,token_type_ids=A__ ,labels=A__) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels)) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__): lowercase = self.num_choices lowercase = NystromformerForMultipleChoice(config=A__) model.to(A__) model.eval() lowercase = input_ids.unsqueeze(1).expand(-1 ,self.num_choices ,-1).contiguous() lowercase = token_type_ids.unsqueeze(1).expand(-1 ,self.num_choices ,-1).contiguous() lowercase = input_mask.unsqueeze(1).expand(-1 ,self.num_choices ,-1).contiguous() lowercase = model( A__ ,attention_mask=A__ ,token_type_ids=A__ ,labels=A__ ,) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices)) def A__ ( self): lowercase = self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) = config_and_inputs lowercase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class lowercase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): lowercase_ : List[str] =( ( NystromformerModel, NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, ) if is_torch_available() else () ) lowercase_ : Optional[int] =( { '''feature-extraction''': NystromformerModel, '''fill-mask''': NystromformerForMaskedLM, '''question-answering''': NystromformerForQuestionAnswering, '''text-classification''': NystromformerForSequenceClassification, '''token-classification''': NystromformerForTokenClassification, '''zero-shot''': NystromformerForSequenceClassification, } if is_torch_available() else {} ) lowercase_ : Any =False lowercase_ : Dict =False def A__ ( self): lowercase = NystromformerModelTester(self) lowercase = ConfigTester(self ,config_class=A__ ,hidden_size=3_7) def A__ ( self): self.config_tester.run_common_tests() def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowercase = type self.model_tester.create_and_check_model(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*A__) @slow def A__ ( self): for model_name in NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase = NystromformerModel.from_pretrained(A__) self.assertIsNotNone(A__) @require_torch class lowercase ( unittest.TestCase ): @slow def A__ ( self): lowercase = NystromformerModel.from_pretrained('''uw-madison/nystromformer-512''') lowercase = torch.tensor([[0, 1, 2, 3, 4, 5]]) with torch.no_grad(): lowercase = model(A__)[0] lowercase = torch.Size((1, 6, 7_6_8)) self.assertEqual(output.shape ,A__) lowercase = torch.tensor( [[[-0.4532, -0.0936, 0.5137], [-0.2676, 0.0628, 0.6186], [-0.3629, -0.1726, 0.4716]]]) self.assertTrue(torch.allclose(output[:, :3, :3] ,A__ ,atol=1E-4)) @slow def A__ ( self): lowercase = '''the [MASK] of Belgium is Brussels''' lowercase = AutoTokenizer.from_pretrained('''uw-madison/nystromformer-512''') lowercase = NystromformerForMaskedLM.from_pretrained('''uw-madison/nystromformer-512''') lowercase = tokenizer(A__ ,return_tensors='''pt''') with torch.no_grad(): lowercase = model(encoding.input_ids).logits lowercase = token_logits[:, 2, :].argmax(-1)[0] self.assertEqual(tokenizer.decode(A__) ,'''capital''')
633
import math import time from typing import Dict, List, Optional from torch.utils.data import Dataset from transformers import SeqaSeqTrainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class lowercase ( SCREAMING_SNAKE_CASE__ ): def __init__( self ,*A__ ,A__=None ,A__=None ,**A__): super().__init__(*A__ ,**A__) lowercase = eval_examples lowercase = post_process_function def A__ ( self ,A__ = None ,A__=None ,A__ = None ,A__ = "eval" ,**A__ ,): lowercase = gen_kwargs.copy() lowercase = ( gen_kwargs['''max_length'''] if gen_kwargs.get('''max_length''') is not None else self.args.generation_max_length ) lowercase = ( gen_kwargs['''num_beams'''] if gen_kwargs.get('''num_beams''') is not None else self.args.generation_num_beams ) lowercase = gen_kwargs lowercase = self.eval_dataset if eval_dataset is None else eval_dataset lowercase = self.get_eval_dataloader(A__) lowercase = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. lowercase = self.compute_metrics lowercase = None lowercase = time.time() lowercase = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: lowercase = eval_loop( A__ ,description='''Evaluation''' ,prediction_loss_only=True if compute_metrics is None else None ,ignore_keys=A__ ,metric_key_prefix=A__ ,) finally: lowercase = compute_metrics lowercase = self.args.eval_batch_size * self.args.world_size if f'{metric_key_prefix}_jit_compilation_time' in output.metrics: start_time += output.metrics[f'{metric_key_prefix}_jit_compilation_time'] output.metrics.update( speed_metrics( A__ ,A__ ,num_samples=output.num_samples ,num_steps=math.ceil(output.num_samples / total_batch_size) ,)) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default lowercase = self.post_process_function(A__ ,A__ ,A__) lowercase = self.compute_metrics(A__) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys()): if not key.startswith(f'{metric_key_prefix}_'): lowercase = metrics.pop(A__) metrics.update(output.metrics) else: lowercase = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(A__) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report()) lowercase = self.callback_handler.on_evaluate(self.args ,self.state ,self.control ,A__) return metrics def A__ ( self ,A__ ,A__ ,A__=None ,A__ = "test" ,**A__): lowercase = gen_kwargs.copy() lowercase = self.get_test_dataloader(A__) # Temporarily disable metric computation, we will do it in the loop here. lowercase = self.compute_metrics lowercase = None lowercase = time.time() lowercase = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: lowercase = eval_loop( A__ ,description='''Prediction''' ,prediction_loss_only=True if compute_metrics is None else None ,ignore_keys=A__ ,metric_key_prefix=A__ ,) finally: lowercase = compute_metrics lowercase = self.args.eval_batch_size * self.args.world_size if f'{metric_key_prefix}_jit_compilation_time' in output.metrics: start_time += output.metrics[f'{metric_key_prefix}_jit_compilation_time'] output.metrics.update( speed_metrics( A__ ,A__ ,num_samples=output.num_samples ,num_steps=math.ceil(output.num_samples / total_batch_size) ,)) if self.post_process_function is None or self.compute_metrics is None: return output lowercase = self.post_process_function(A__ ,A__ ,A__ ,'''predict''') lowercase = self.compute_metrics(A__) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys()): if not key.startswith(f'{metric_key_prefix}_'): lowercase = metrics.pop(A__) metrics.update(output.metrics) return PredictionOutput(predictions=predictions.predictions ,label_ids=predictions.label_ids ,metrics=A__)
633
1
import math import sys import cva import numpy as np def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' # For applying gaussian function for each element in matrix. lowercase = math.sqrt(lowerCAmelCase__ ) lowercase = 1 / (sigma * math.sqrt(2 * math.pi )) return cons * np.exp(-((img / sigma) ** 2) * 0.5 ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = kernel_size // 2 return img[x - half : x + half + 1, y - half : y + half + 1] def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' # Creates a gaussian kernel of given dimension. lowercase = np.zeros((kernel_size, kernel_size) ) for i in range(0 , lowerCAmelCase__ ): for j in range(0 , lowerCAmelCase__ ): lowercase = math.sqrt( abs(i - kernel_size // 2 ) ** 2 + abs(j - kernel_size // 2 ) ** 2 ) return vec_gaussian(lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ): '''simple docstring''' lowercase = np.zeros(img.shape ) lowercase = get_gauss_kernel(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase , lowercase = img.shape for i in range(kernel_size // 2 , size_x - kernel_size // 2 ): for j in range(kernel_size // 2 , size_y - kernel_size // 2 ): lowercase = get_slice(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) lowercase = img_s - img_s[kernel_size // 2, kernel_size // 2] lowercase = vec_gaussian(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase = np.multiply(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase = np.multiply(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase = np.sum(lowerCAmelCase__ ) / np.sum(lowerCAmelCase__ ) lowercase = val return imga def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = args[1] if args[1:] else '''../image_data/lena.jpg''' lowercase = float(args[2] ) if args[2:] else 1.0 lowercase = float(args[3] ) if args[3:] else 1.0 if args[4:]: lowercase = int(args[4] ) lowercase = kernel_size + abs(kernel_size % 2 - 1 ) else: lowercase = 5 return filename, spatial_variance, intensity_variance, kernel_size if __name__ == "__main__": lowercase__ , lowercase__ , lowercase__ , lowercase__ :List[str] = parse_args(sys.argv) lowercase__ :Union[str, Any] = cva.imread(filename, 0) cva.imshow("input image", img) lowercase__ :Tuple = img / 255 lowercase__ :Tuple = out.astype("float32") lowercase__ :str = bilateral_filter(out, spatial_variance, intensity_variance, kernel_size) lowercase__ :Optional[int] = out * 255 lowercase__ :str = np.uinta(out) cva.imshow("output image", out) cva.waitKey(0) cva.destroyAllWindows()
633
# Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import subprocess from packaging.version import Version, parse from accelerate.commands.config.config_args import default_config_file, load_config_from_file lowercase__ :Any = "Run commands across TPU VMs for initial setup before running `accelerate launch`." def UpperCamelCase ( lowerCAmelCase__=None ): '''simple docstring''' if subparsers is not None: lowercase = subparsers.add_parser('''tpu-config''' , description=_description ) else: lowercase = argparse.ArgumentParser('''Accelerate tpu-config command''' , description=_description ) # Core arguments lowercase = parser.add_argument_group( '''Config Arguments''' , '''Arguments that can be configured through `accelerate config`.''' ) config_args.add_argument( '''--config_file''' , type=lowerCAmelCase__ , default=lowerCAmelCase__ , help='''Path to the config file to use for accelerate.''' , ) config_args.add_argument( '''--tpu_name''' , default=lowerCAmelCase__ , help='''The name of the TPU to use. If not specified, will use the TPU specified in the config file.''' , ) config_args.add_argument( '''--tpu_zone''' , default=lowerCAmelCase__ , help='''The zone of the TPU to use. If not specified, will use the zone specified in the config file.''' , ) lowercase = parser.add_argument_group('''TPU Arguments''' , '''Arguments for options ran inside the TPU.''' ) pod_args.add_argument( '''--use_alpha''' , action='''store_true''' , help='''Whether to use `gcloud alpha` when running the TPU training script instead of `gcloud`.''' , ) pod_args.add_argument( '''--command_file''' , default=lowerCAmelCase__ , help='''The path to the file containing the commands to run on the pod on startup.''' , ) pod_args.add_argument( '''--command''' , action='''append''' , nargs='''+''' , help='''A command to run on the pod. Can be passed multiple times.''' , ) pod_args.add_argument( '''--install_accelerate''' , action='''store_true''' , help='''Whether to install accelerate on the pod. Defaults to False.''' , ) pod_args.add_argument( '''--accelerate_version''' , default='''latest''' , help='''The version of accelerate to install on the pod. If not specified, will use the latest pypi version. Specify \'dev\' to install from GitHub.''' , ) pod_args.add_argument( '''--debug''' , action='''store_true''' , help='''If set, will print the command that would be run instead of running it.''' ) if subparsers is not None: parser.set_defaults(func=lowerCAmelCase__ ) return parser def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = None # Get the default from the config file if it exists. if args.config_file is not None or os.path.isfile(lowerCAmelCase__ ): lowercase = load_config_from_file(args.config_file ) if not args.command_file and defaults.command_file is not None and not args.command: lowercase = defaults.command_file if not args.command and defaults.commands is not None: lowercase = defaults.commands if not args.tpu_name: lowercase = defaults.tpu_name if not args.tpu_zone: lowercase = defaults.tpu_zone if args.accelerate_version == "dev": lowercase = '''git+https://github.com/huggingface/accelerate.git''' elif args.accelerate_version == "latest": lowercase = '''accelerate -U''' elif isinstance(parse(args.accelerate_version ) , lowerCAmelCase__ ): lowercase = f'accelerate=={args.accelerate_version}' if not args.command_file and not args.command: raise ValueError('''You must specify either a command file or a command to run on the pod.''' ) if args.command_file: with open(args.command_file , '''r''' ) as f: lowercase = [f.read().splitlines()] # To turn list of lists into list of strings if isinstance(args.command[0] , lowerCAmelCase__ ): lowercase = [line for cmd in args.command for line in cmd] # Default to the shared folder and install accelerate lowercase = ['''cd /usr/share'''] if args.install_accelerate: new_cmd += [f'pip install {args.accelerate_version}'] new_cmd += args.command lowercase = '''; '''.join(lowerCAmelCase__ ) # Then send it to gcloud # Eventually try to use google-api-core to do this instead of subprocess lowercase = ['''gcloud'''] if args.use_alpha: cmd += ["alpha"] cmd += [ "compute", "tpus", "tpu-vm", "ssh", args.tpu_name, "--zone", args.tpu_zone, "--command", args.command, "--worker", "all", ] if args.debug: print(f'Running {" ".join(lowerCAmelCase__ )}' ) return subprocess.run(lowerCAmelCase__ ) print('''Successfully setup pod.''' ) def UpperCamelCase ( ): '''simple docstring''' lowercase = tpu_command_parser() lowercase = parser.parse_args() tpu_command_launcher(lowerCAmelCase__ )
633
1
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ :List[str] = logging.get_logger(__name__) lowercase__ :List[str] = { "BridgeTower/bridgetower-base": "https://huggingface.co/BridgeTower/bridgetower-base/blob/main/config.json", "BridgeTower/bridgetower-base-itm-mlm": ( "https://huggingface.co/BridgeTower/bridgetower-base-itm-mlm/blob/main/config.json" ), } class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Optional[int] ='''bridgetower_vision_model''' def __init__( self ,A__=7_6_8 ,A__=1_2 ,A__=3 ,A__=1_6 ,A__=2_8_8 ,A__=1 ,A__=1E-05 ,A__=False ,A__=True ,A__=False ,**A__ ,): super().__init__(**A__) lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_channels lowercase = patch_size lowercase = image_size lowercase = initializer_factor lowercase = layer_norm_eps lowercase = stop_gradient lowercase = share_layernorm lowercase = remove_last_layer @classmethod def A__ ( cls ,A__ ,**A__): lowercase , lowercase = cls.get_config_dict(A__ ,**A__) if config_dict.get('''model_type''') == "bridgetower": lowercase = config_dict['''text_config'''] if "model_type" in config_dict and hasattr(cls ,'''model_type''') and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.') return cls.from_dict(A__ ,**A__) class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Optional[int] ='''bridgetower_text_model''' def __init__( self ,A__=5_0_2_6_5 ,A__=7_6_8 ,A__=1_2 ,A__=1_2 ,A__=1 ,A__=3_0_7_2 ,A__="gelu" ,A__=0.1 ,A__=0.1 ,A__=5_1_4 ,A__=1 ,A__=1E-05 ,A__=1 ,A__=0 ,A__=2 ,A__="absolute" ,A__=True ,**A__ ,): super().__init__(**A__) lowercase = vocab_size lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_attention_heads lowercase = hidden_act lowercase = initializer_factor lowercase = intermediate_size lowercase = hidden_dropout_prob lowercase = attention_probs_dropout_prob lowercase = max_position_embeddings lowercase = type_vocab_size lowercase = layer_norm_eps lowercase = position_embedding_type lowercase = use_cache lowercase = pad_token_id lowercase = bos_token_id lowercase = eos_token_id @classmethod def A__ ( cls ,A__ ,**A__): lowercase , lowercase = cls.get_config_dict(A__ ,**A__) if config_dict.get('''model_type''') == "bridgetower": lowercase = config_dict['''text_config'''] if "model_type" in config_dict and hasattr(cls ,'''model_type''') and config_dict["model_type"] != cls.model_type: logger.warning( f'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.') return cls.from_dict(A__ ,**A__) class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Any ='''bridgetower''' def __init__( self ,A__=True ,A__="gelu" ,A__=7_6_8 ,A__=1 ,A__=1E-05 ,A__=False ,A__="add" ,A__=1_2 ,A__=6 ,A__=False ,A__=False ,A__=None ,A__=None ,**A__ ,): # TODO: remove this once the Hub files are updated. lowercase = kwargs.pop('''text_config_dict''' ,A__) lowercase = kwargs.pop('''vision_config_dict''' ,A__) super().__init__(**A__) lowercase = share_cross_modal_transformer_layers lowercase = hidden_act lowercase = hidden_size lowercase = initializer_factor lowercase = layer_norm_eps lowercase = share_link_tower_layers lowercase = link_tower_type lowercase = num_attention_heads lowercase = num_hidden_layers lowercase = tie_word_embeddings lowercase = init_layernorm_from_vision_encoder if text_config is None: lowercase = {} logger.info('''`text_config` is `None`. Initializing the `BridgeTowerTextConfig` with default values.''') if vision_config is None: lowercase = {} logger.info('''`vision_config` is `None`. Initializing the `BridgeTowerVisionConfig` with default values.''') lowercase = BridgeTowerTextConfig(**A__) lowercase = BridgeTowerVisionConfig(**A__) @classmethod def A__ ( cls ,A__ ,A__ ,**A__): return cls(text_config=text_config.to_dict() ,vision_config=vision_config.to_dict() ,**A__) def A__ ( self): lowercase = copy.deepcopy(self.__dict__) lowercase = self.text_config.to_dict() lowercase = self.vision_config.to_dict() lowercase = self.__class__.model_type return output
633
import json import multiprocessing as mp import re from collections import defaultdict from functools import partial from typing import Dict, List, Optional, Set, Tuple, Type from datasets import Dataset from datasketch import MinHash, MinHashLSH from dpu_utils.utils.iterators import ThreadedIterator from tqdm import tqdm lowercase__ :Optional[Any] = re.compile("[^A-Za-z_0-9]") # parameters used in DuplicationIndex lowercase__ :List[str] = 10 lowercase__ :Tuple = 256 def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if len(lowerCAmelCase__ ) < MIN_NUM_TOKENS: return None lowercase = MinHash(num_perm=lowerCAmelCase__ ) for token in set(lowerCAmelCase__ ): min_hash.update(token.encode() ) return min_hash def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return {t for t in NON_ALPHA.split(lowerCAmelCase__ ) if len(t.strip() ) > 0} class lowercase : def __init__( self ,*, A__ = 0.85 ,): lowercase = duplication_jaccard_threshold lowercase = NUM_PERM lowercase = MinHashLSH(threshold=self._duplication_jaccard_threshold ,num_perm=self._num_perm) lowercase = defaultdict(A__) def A__ ( self ,A__ ,A__): lowercase = self._index.query(A__) if code_key in self._index.keys: print(f'Duplicate key {code_key}') return self._index.insert(A__ ,A__) if len(A__) > 0: for base_duplicate in close_duplicates: if base_duplicate in self._duplicate_clusters: self._duplicate_clusters[base_duplicate].add(A__) break else: self._duplicate_clusters[close_duplicates[0]].add(A__) def A__ ( self): lowercase = [] for base, duplicates in self._duplicate_clusters.items(): lowercase = [base] + list(A__) # reformat the cluster to be a list of dict lowercase = [{'''base_index''': el[0], '''repo_name''': el[1], '''path''': el[2]} for el in cluster] duplicate_clusters.append(A__) return duplicate_clusters def A__ ( self ,A__): lowercase = self.get_duplicate_clusters() with open(A__ ,'''w''') as f: json.dump(A__ ,A__) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase = element lowercase = get_min_hash([t for t in NON_ALPHA.split(data['''content'''] ) if len(t.strip() ) > 0] ) if min_hash is not None: return (index, data["repo_name"], data["path"]), min_hash def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' with mp.Pool() as pool: for data in pool.imap_unordered( _compute_min_hash , ThreadedIterator(lowerCAmelCase__ , max_queue_size=1_0000 ) , chunksize=100 , ): if data is not None: yield data def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = DuplicationIndex(duplication_jaccard_threshold=lowerCAmelCase__ ) for filename, min_hash in tqdm(ThreadedIterator(minhash_iter(enumerate(lowerCAmelCase__ ) ) , max_queue_size=100 ) ): di.add(lowerCAmelCase__ , lowerCAmelCase__ ) # Returns a List[Cluster] where Cluster is List[str] with the filenames. return di.get_duplicate_clusters() def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = get_tokens(lowerCAmelCase__ ) lowercase = get_tokens(lowerCAmelCase__ ) return len(tokensa & tokensa ) / len(tokensa | tokensa ) lowercase__ :List[Any] = None def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = [] for elementa in cluster: lowercase = _shared_dataset[elementa['''base_index''']]['''content'''] for elementa in extremes: lowercase = _shared_dataset[elementa['''base_index''']]['''content'''] if jaccard_similarity(lowerCAmelCase__ , lowerCAmelCase__ ) >= jaccard_threshold: elementa["copies"] += 1 break else: lowercase = 1 extremes.append(lowerCAmelCase__ ) return extremes def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' global _shared_dataset lowercase = dataset lowercase = [] lowercase = partial(_find_cluster_extremes_shared , jaccard_threshold=lowerCAmelCase__ ) with mp.Pool() as pool: for extremes in tqdm( pool.imap_unordered( lowerCAmelCase__ , lowerCAmelCase__ , ) , total=len(lowerCAmelCase__ ) , ): extremes_list.append(lowerCAmelCase__ ) return extremes_list def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ = 0.85 ): '''simple docstring''' lowercase = make_duplicate_clusters(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase = {x['''base_index'''] for cluster in duplicate_clusters for x in cluster} lowercase = {} lowercase = find_extremes(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) for extremes in extremes_clusters: for element in extremes: lowercase = element lowercase = duplicate_indices - set(extreme_dict.keys() ) lowercase = dataset.filter(lambda lowerCAmelCase__ , lowerCAmelCase__ : idx not in remove_indices , with_indices=lowerCAmelCase__ ) # update duplicate_clusters for cluster in duplicate_clusters: for element in cluster: lowercase = element['''base_index'''] in extreme_dict if element["is_extreme"]: lowercase = extreme_dict[element['''base_index''']]['''copies'''] print(f'Original dataset size: {len(lowerCAmelCase__ )}' ) print(f'Number of duplicate clusters: {len(lowerCAmelCase__ )}' ) print(f'Files in duplicate cluster: {len(lowerCAmelCase__ )}' ) print(f'Unique files in duplicate cluster: {len(lowerCAmelCase__ )}' ) print(f'Filtered dataset size: {len(lowerCAmelCase__ )}' ) return ds_filter, duplicate_clusters
633
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase__ :str = logging.get_logger(__name__) lowercase__ :List[Any] = { "distilbert-base-uncased": "https://huggingface.co/distilbert-base-uncased/resolve/main/config.json", "distilbert-base-uncased-distilled-squad": ( "https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json" ), "distilbert-base-cased": "https://huggingface.co/distilbert-base-cased/resolve/main/config.json", "distilbert-base-cased-distilled-squad": ( "https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json" ), "distilbert-base-german-cased": "https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json", "distilbert-base-multilingual-cased": ( "https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json" ), "distilbert-base-uncased-finetuned-sst-2-english": ( "https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json" ), } class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : str ='''distilbert''' lowercase_ : Union[str, Any] ={ '''hidden_size''': '''dim''', '''num_attention_heads''': '''n_heads''', '''num_hidden_layers''': '''n_layers''', } def __init__( self ,A__=3_0_5_2_2 ,A__=5_1_2 ,A__=False ,A__=6 ,A__=1_2 ,A__=7_6_8 ,A__=4 * 7_6_8 ,A__=0.1 ,A__=0.1 ,A__="gelu" ,A__=0.02 ,A__=0.1 ,A__=0.2 ,A__=0 ,**A__ ,): lowercase = vocab_size lowercase = max_position_embeddings lowercase = sinusoidal_pos_embds lowercase = n_layers lowercase = n_heads lowercase = dim lowercase = hidden_dim lowercase = dropout lowercase = attention_dropout lowercase = activation lowercase = initializer_range lowercase = qa_dropout lowercase = seq_classif_dropout super().__init__(**A__ ,pad_token_id=A__) class lowercase ( SCREAMING_SNAKE_CASE__ ): @property def A__ ( self): if self.task == "multiple-choice": lowercase = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: lowercase = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ])
633
import numpy as np import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModelWithProjection, PreTrainedModel from ...utils import logging lowercase__ :Union[str, Any] = logging.get_logger(__name__) class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Union[str, Any] =CLIPConfig lowercase_ : str =['''CLIPEncoderLayer'''] def __init__( self ,A__): super().__init__(A__) lowercase = CLIPVisionModelWithProjection(config.vision_config) lowercase = nn.Linear(config.vision_config.projection_dim ,1) lowercase = nn.Linear(config.vision_config.projection_dim ,1) @torch.no_grad() def A__ ( self ,A__ ,A__ ,A__=0.5 ,A__=0.5): lowercase = self.vision_model(A__)[0] lowercase = self.p_head(A__) lowercase = nsfw_detected.flatten() lowercase = nsfw_detected > p_threshold lowercase = nsfw_detected.tolist() if any(A__): logger.warning( '''Potential NSFW content was detected in one or more images. A black image will be returned instead.''' ''' Try again with a different prompt and/or seed.''') for idx, nsfw_detected_ in enumerate(A__): if nsfw_detected_: lowercase = np.zeros(images[idx].shape) lowercase = self.w_head(A__) lowercase = watermark_detected.flatten() lowercase = watermark_detected > w_threshold lowercase = watermark_detected.tolist() if any(A__): logger.warning( '''Potential watermarked content was detected in one or more images. A black image will be returned instead.''' ''' Try again with a different prompt and/or seed.''') for idx, watermark_detected_ in enumerate(A__): if watermark_detected_: lowercase = np.zeros(images[idx].shape) return images, nsfw_detected, watermark_detected
633
1
import inspect import unittest from transformers import RegNetConfig, is_flax_available from transformers.testing_utils import require_flax, slow from transformers.utils import cached_property, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.models.regnet.modeling_flax_regnet import FlaxRegNetForImageClassification, FlaxRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowercase ( unittest.TestCase ): def __init__( self ,A__ ,A__=3 ,A__=3_2 ,A__=3 ,A__=1_0 ,A__=[1_0, 2_0, 3_0, 4_0] ,A__=[1, 1, 2, 1] ,A__=True ,A__=True ,A__="relu" ,A__=3 ,A__=None ,): lowercase = parent lowercase = batch_size lowercase = image_size lowercase = num_channels lowercase = embeddings_size lowercase = hidden_sizes lowercase = depths lowercase = is_training lowercase = use_labels lowercase = hidden_act lowercase = num_labels lowercase = scope lowercase = len(A__) def A__ ( self): lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) lowercase = self.get_config() return config, pixel_values def A__ ( self): return RegNetConfig( num_channels=self.num_channels ,embeddings_size=self.embeddings_size ,hidden_sizes=self.hidden_sizes ,depths=self.depths ,hidden_act=self.hidden_act ,num_labels=self.num_labels ,image_size=self.image_size ,) def A__ ( self ,A__ ,A__): lowercase = FlaxRegNetModel(config=A__) lowercase = model(A__) # Output shape (b, c, h, w) self.parent.assertEqual( result.last_hidden_state.shape ,(self.batch_size, self.hidden_sizes[-1], self.image_size // 3_2, self.image_size // 3_2) ,) def A__ ( self ,A__ ,A__): lowercase = self.num_labels lowercase = FlaxRegNetForImageClassification(config=A__) lowercase = model(A__) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels)) def A__ ( self): lowercase = self.prepare_config_and_inputs() lowercase , lowercase = config_and_inputs lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_flax class lowercase ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): lowercase_ : Union[str, Any] =(FlaxRegNetModel, FlaxRegNetForImageClassification) if is_flax_available() else () lowercase_ : Optional[int] =False lowercase_ : str =False lowercase_ : Any =False def A__ ( self): lowercase = FlaxRegNetModelTester(self) lowercase = ConfigTester(self ,config_class=A__ ,has_text_modality=A__) def A__ ( self): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def A__ ( self): return def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A__) @unittest.skip(reason='''RegNet does not use inputs_embeds''') def A__ ( self): pass @unittest.skip(reason='''RegNet does not support input and output embeddings''') def A__ ( self): pass def A__ ( self): lowercase , lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase = model_class(A__) lowercase = inspect.signature(model.__call__) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase = [*signature.parameters.keys()] lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,A__) def A__ ( self): def check_hidden_states_output(A__ ,A__ ,A__): lowercase = model_class(A__) lowercase = model(**self._prepare_for_class(A__ ,A__)) lowercase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowercase = self.model_tester.num_stages self.assertEqual(len(A__) ,expected_num_stages + 1) lowercase , lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase = True check_hidden_states_output(A__ ,A__ ,A__) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowercase = True check_hidden_states_output(A__ ,A__ ,A__) def A__ ( self): lowercase , lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__): lowercase = self._prepare_for_class(A__ ,A__) lowercase = model_class(A__) @jax.jit def model_jitted(A__ ,**A__): return model(pixel_values=A__ ,**A__) with self.subTest('''JIT Enabled'''): lowercase = model_jitted(**A__).to_tuple() with self.subTest('''JIT Disabled'''): with jax.disable_jit(): lowercase = model_jitted(**A__).to_tuple() self.assertEqual(len(A__) ,len(A__)) for jitted_output, output in zip(A__ ,A__): self.assertEqual(jitted_output.shape ,output.shape) def UpperCamelCase ( ): '''simple docstring''' lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_flax class lowercase ( unittest.TestCase ): @cached_property def A__ ( self): return AutoImageProcessor.from_pretrained('''facebook/regnet-y-040''') if is_vision_available() else None @slow def A__ ( self): lowercase = FlaxRegNetForImageClassification.from_pretrained('''facebook/regnet-y-040''') lowercase = self.default_image_processor lowercase = prepare_img() lowercase = image_processor(images=A__ ,return_tensors='''np''') lowercase = model(**A__) # verify the logits lowercase = (1, 1_0_0_0) self.assertEqual(outputs.logits.shape ,A__) lowercase = jnp.array([-0.4180, -1.5051, -3.4836]) self.assertTrue(jnp.allclose(outputs.logits[0, :3] ,A__ ,atol=1E-4))
633
class lowercase : def __init__( self ,A__): lowercase = val lowercase = None lowercase = None def A__ ( self ,A__): if self.val: if val < self.val: if self.left is None: lowercase = Node(A__) else: self.left.insert(A__) elif val > self.val: if self.right is None: lowercase = Node(A__) else: self.right.insert(A__) else: lowercase = val def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' # Recursive traversal if root: inorder(root.left , lowerCAmelCase__ ) res.append(root.val ) inorder(root.right , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' # Build BST if len(lowerCAmelCase__ ) == 0: return arr lowercase = Node(arr[0] ) for i in range(1 , len(lowerCAmelCase__ ) ): root.insert(arr[i] ) # Traverse BST in order. lowercase = [] inorder(lowerCAmelCase__ , lowerCAmelCase__ ) return res if __name__ == "__main__": print(tree_sort([10, 1, 3, 2, 9, 14, 13]))
633
1
from typing import TYPE_CHECKING from ...utils import _LazyModule lowercase__ :List[Any] = {"tokenization_bertweet": ["BertweetTokenizer"]} if TYPE_CHECKING: from .tokenization_bertweet import BertweetTokenizer else: import sys lowercase__ :Optional[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
633
import os def UpperCamelCase ( lowerCAmelCase__ = "input.txt" ): '''simple docstring''' with open(os.path.join(os.path.dirname(lowerCAmelCase__ ) , lowerCAmelCase__ ) ) as input_file: lowercase = [ [int(lowerCAmelCase__ ) for element in line.split(''',''' )] for line in input_file.readlines() ] lowercase = len(lowerCAmelCase__ ) lowercase = len(matrix[0] ) lowercase = [[-1 for _ in range(lowerCAmelCase__ )] for _ in range(lowerCAmelCase__ )] for i in range(lowerCAmelCase__ ): lowercase = matrix[i][0] for j in range(1 , lowerCAmelCase__ ): for i in range(lowerCAmelCase__ ): lowercase = minimal_path_sums[i][j - 1] + matrix[i][j] for i in range(1 , lowerCAmelCase__ ): lowercase = min( minimal_path_sums[i][j] , minimal_path_sums[i - 1][j] + matrix[i][j] ) for i in range(rows - 2 , -1 , -1 ): lowercase = min( minimal_path_sums[i][j] , minimal_path_sums[i + 1][j] + matrix[i][j] ) return min(minimal_path_sums_row[-1] for minimal_path_sums_row in minimal_path_sums ) if __name__ == "__main__": print(F'{solution() = }')
633
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ :Any = { "configuration_x_clip": [ "XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "XCLIPConfig", "XCLIPTextConfig", "XCLIPVisionConfig", ], "processing_x_clip": ["XCLIPProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :Union[str, Any] = [ "XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "XCLIPModel", "XCLIPPreTrainedModel", "XCLIPTextModel", "XCLIPVisionModel", ] if TYPE_CHECKING: from .configuration_x_clip import ( XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, XCLIPConfig, XCLIPTextConfig, XCLIPVisionConfig, ) from .processing_x_clip import XCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_x_clip import ( XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, XCLIPModel, XCLIPPreTrainedModel, XCLIPTextModel, XCLIPVisionModel, ) else: import sys lowercase__ :Tuple = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
633
from __future__ import annotations def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if len(lowerCAmelCase__ ) < 2: raise ValueError('''Monogons and Digons are not polygons in the Euclidean space''' ) if any(i <= 0 for i in nums ): raise ValueError('''All values must be greater than 0''' ) lowercase = nums.copy() copy_nums.sort() return copy_nums[-1] < sum(copy_nums[:-1] ) if __name__ == "__main__": import doctest doctest.testmod()
633
1
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_mbart import MBartTokenizer else: lowercase__ :List[Any] = None lowercase__ :Any = logging.get_logger(__name__) lowercase__ :List[Any] = {"vocab_file": "sentencepiece.bpe.model", "tokenizer_file": "tokenizer.json"} lowercase__ :List[Any] = { "vocab_file": { "facebook/mbart-large-en-ro": ( "https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model" ), "facebook/mbart-large-cc25": ( "https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model" ), }, "tokenizer_file": { "facebook/mbart-large-en-ro": "https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/tokenizer.json", "facebook/mbart-large-cc25": "https://huggingface.co/facebook/mbart-large-cc25/resolve/main/tokenizer.json", }, } lowercase__ :List[Any] = { "facebook/mbart-large-en-ro": 1024, "facebook/mbart-large-cc25": 1024, } # fmt: off lowercase__ :Union[str, Any] = ["ar_AR", "cs_CZ", "de_DE", "en_XX", "es_XX", "et_EE", "fi_FI", "fr_XX", "gu_IN", "hi_IN", "it_IT", "ja_XX", "kk_KZ", "ko_KR", "lt_LT", "lv_LV", "my_MM", "ne_NP", "nl_XX", "ro_RO", "ru_RU", "si_LK", "tr_TR", "vi_VN", "zh_CN"] class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Tuple =VOCAB_FILES_NAMES lowercase_ : List[Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ : List[str] =PRETRAINED_VOCAB_FILES_MAP lowercase_ : Optional[Any] =['''input_ids''', '''attention_mask'''] lowercase_ : Optional[int] =MBartTokenizer lowercase_ : List[int] =[] lowercase_ : List[int] =[] def __init__( self ,A__=None ,A__=None ,A__="<s>" ,A__="</s>" ,A__="</s>" ,A__="<s>" ,A__="<unk>" ,A__="<pad>" ,A__="<mask>" ,A__=None ,A__=None ,A__=None ,**A__ ,): # Mask token behave like a normal word, i.e. include the space before it lowercase = AddedToken(A__ ,lstrip=A__ ,rstrip=A__) if isinstance(A__ ,A__) else mask_token super().__init__( vocab_file=A__ ,tokenizer_file=A__ ,bos_token=A__ ,eos_token=A__ ,sep_token=A__ ,cls_token=A__ ,unk_token=A__ ,pad_token=A__ ,mask_token=A__ ,src_lang=A__ ,tgt_lang=A__ ,additional_special_tokens=A__ ,**A__ ,) lowercase = vocab_file lowercase = False if not self.vocab_file else True lowercase = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens]) self.add_special_tokens({'''additional_special_tokens''': _additional_special_tokens}) lowercase = { lang_code: self.convert_tokens_to_ids(A__) for lang_code in FAIRSEQ_LANGUAGE_CODES } lowercase = src_lang if src_lang is not None else '''en_XX''' lowercase = self.convert_tokens_to_ids(self._src_lang) lowercase = tgt_lang self.set_src_lang_special_tokens(self._src_lang) @property def A__ ( self): return self._src_lang @src_lang.setter def A__ ( self ,A__): lowercase = new_src_lang self.set_src_lang_special_tokens(self._src_lang) def A__ ( self ,A__ ,A__ = None): if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def A__ ( self ,A__ ,A__ = None): lowercase = [self.sep_token_id] lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] def A__ ( self ,A__ ,A__ ,A__ ,A__ ,**A__): if src_lang is None or tgt_lang is None: raise ValueError('''Translation requires a `src_lang` and a `tgt_lang` for this model''') lowercase = src_lang lowercase = self(A__ ,add_special_tokens=A__ ,return_tensors=A__ ,**A__) lowercase = self.convert_tokens_to_ids(A__) lowercase = tgt_lang_id return inputs def A__ ( self ,A__ ,A__ = "en_XX" ,A__ = None ,A__ = "ro_RO" ,**A__ ,): lowercase = src_lang lowercase = tgt_lang return super().prepare_seqaseq_batch(A__ ,A__ ,**A__) def A__ ( self): return self.set_src_lang_special_tokens(self.src_lang) def A__ ( self): return self.set_tgt_lang_special_tokens(self.tgt_lang) def A__ ( self ,A__): lowercase = self.convert_tokens_to_ids(A__) lowercase = [] lowercase = [self.eos_token_id, self.cur_lang_code] lowercase = self.convert_ids_to_tokens(self.prefix_tokens) lowercase = self.convert_ids_to_tokens(self.suffix_tokens) lowercase = processors.TemplateProcessing( single=prefix_tokens_str + ['''$A'''] + suffix_tokens_str ,pair=prefix_tokens_str + ['''$A''', '''$B'''] + suffix_tokens_str ,special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str ,self.prefix_tokens + self.suffix_tokens)) ,) def A__ ( self ,A__): lowercase = self.convert_tokens_to_ids(A__) lowercase = [] lowercase = [self.eos_token_id, self.cur_lang_code] lowercase = self.convert_ids_to_tokens(self.prefix_tokens) lowercase = self.convert_ids_to_tokens(self.suffix_tokens) lowercase = processors.TemplateProcessing( single=prefix_tokens_str + ['''$A'''] + suffix_tokens_str ,pair=prefix_tokens_str + ['''$A''', '''$B'''] + suffix_tokens_str ,special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str ,self.prefix_tokens + self.suffix_tokens)) ,) def A__ ( self ,A__ ,A__ = None): if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''') if not os.path.isdir(A__): logger.error(f'Vocabulary path ({save_directory}) should be a directory.') return lowercase = os.path.join( A__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file''']) if os.path.abspath(self.vocab_file) != os.path.abspath(A__): copyfile(self.vocab_file ,A__) return (out_vocab_file,)
633
def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' return int((input_a, input_a).count(0 ) == 0 ) def UpperCamelCase ( ): '''simple docstring''' assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_gate(1 , 0 ) == 0 assert and_gate(1 , 1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
633
1
import json import os import re import unicodedata from json.encoder import INFINITY from typing import Any, Dict, List, Optional, Tuple, Union import numpy as np import regex from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_flax_available, is_tf_available, is_torch_available, logging from ...utils.generic import _is_jax, _is_numpy lowercase__ :List[str] = logging.get_logger(__name__) lowercase__ :Dict = { "artists_file": "artists.json", "lyrics_file": "lyrics.json", "genres_file": "genres.json", } lowercase__ :List[Any] = { "artists_file": { "jukebox": "https://huggingface.co/ArthurZ/jukebox/blob/main/artists.json", }, "genres_file": { "jukebox": "https://huggingface.co/ArthurZ/jukebox/blob/main/genres.json", }, "lyrics_file": { "jukebox": "https://huggingface.co/ArthurZ/jukebox/blob/main/lyrics.json", }, } lowercase__ :Union[str, Any] = { "jukebox": 512, } class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Union[str, Any] =VOCAB_FILES_NAMES lowercase_ : Any =PRETRAINED_VOCAB_FILES_MAP lowercase_ : Tuple =PRETRAINED_LYRIC_TOKENS_SIZES lowercase_ : Tuple =['''input_ids''', '''attention_mask'''] def __init__( self ,A__ ,A__ ,A__ ,A__=["v3", "v2", "v2"] ,A__=5_1_2 ,A__=5 ,A__="<|endoftext|>" ,**A__ ,): lowercase = AddedToken(A__ ,lstrip=A__ ,rstrip=A__) if isinstance(A__ ,A__) else unk_token super().__init__( unk_token=A__ ,n_genres=A__ ,version=A__ ,max_n_lyric_tokens=A__ ,**A__ ,) lowercase = version lowercase = max_n_lyric_tokens lowercase = n_genres with open(A__ ,encoding='''utf-8''') as vocab_handle: lowercase = json.load(A__) with open(A__ ,encoding='''utf-8''') as vocab_handle: lowercase = json.load(A__) with open(A__ ,encoding='''utf-8''') as vocab_handle: lowercase = json.load(A__) lowercase = r'''[^A-Za-z0-9.,:;!?\-\'\"()\[\] \t\n]+''' # In v2, we had a n_vocab=80 and in v3 we missed + and so n_vocab=79 of characters. if len(self.lyrics_encoder) == 7_9: lowercase = oov.replace(r'''\-\'''' ,r'''\-+\'''') lowercase = regex.compile(A__) lowercase = {v: k for k, v in self.artists_encoder.items()} lowercase = {v: k for k, v in self.genres_encoder.items()} lowercase = {v: k for k, v in self.lyrics_encoder.items()} @property def A__ ( self): return len(self.artists_encoder) + len(self.genres_encoder) + len(self.lyrics_encoder) def A__ ( self): return dict(self.artists_encoder ,self.genres_encoder ,self.lyrics_encoder) def A__ ( self ,A__ ,A__ ,A__): lowercase = [self.artists_encoder.get(A__ ,0) for artist in list_artists] for genres in range(len(A__)): lowercase = [self.genres_encoder.get(A__ ,0) for genre in list_genres[genres]] lowercase = list_genres[genres] + [-1] * (self.n_genres - len(list_genres[genres])) lowercase = [[self.lyrics_encoder.get(A__ ,0) for character in list_lyrics[0]], [], []] return artists_id, list_genres, lyric_ids def A__ ( self ,A__): return list(A__) def A__ ( self ,A__ ,A__ ,A__ ,**A__): lowercase , lowercase , lowercase = self.prepare_for_tokenization(A__ ,A__ ,A__) lowercase = self._tokenize(A__) return artist, genre, lyrics def A__ ( self ,A__ ,A__ ,A__ ,A__ = False): for idx in range(len(self.version)): if self.version[idx] == "v3": lowercase = artists[idx].lower() lowercase = [genres[idx].lower()] else: lowercase = self._normalize(artists[idx]) + '''.v2''' lowercase = [ self._normalize(A__) + '''.v2''' for genre in genres[idx].split('''_''') ] # split is for the full dictionary with combined genres if self.version[0] == "v2": lowercase = regex.compile(r'''[^A-Za-z0-9.,:;!?\-\'\"()\[\] \t\n]+''') lowercase = '''ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789.,:;!?-+\'\"()[] \t\n''' lowercase = {vocab[index]: index + 1 for index in range(len(A__))} lowercase = 0 lowercase = len(A__) + 1 lowercase = self.vocab lowercase = {v: k for k, v in self.vocab.items()} lowercase = '''''' else: lowercase = regex.compile(r'''[^A-Za-z0-9.,:;!?\-+\'\"()\[\] \t\n]+''') lowercase = self._run_strip_accents(A__) lowercase = lyrics.replace('''\\''' ,'''\n''') lowercase = self.out_of_vocab.sub('''''' ,A__), [], [] return artists, genres, lyrics def A__ ( self ,A__): lowercase = unicodedata.normalize('''NFD''' ,A__) lowercase = [] for char in text: lowercase = unicodedata.category(A__) if cat == "Mn": continue output.append(A__) return "".join(A__) def A__ ( self ,A__): lowercase = ( [chr(A__) for i in range(ord('''a''') ,ord('''z''') + 1)] + [chr(A__) for i in range(ord('''A''') ,ord('''Z''') + 1)] + [chr(A__) for i in range(ord('''0''') ,ord('''9''') + 1)] + ['''.'''] ) lowercase = frozenset(A__) lowercase = re.compile(r'''_+''') lowercase = ''''''.join([c if c in accepted else '''_''' for c in text.lower()]) lowercase = pattern.sub('''_''' ,A__).strip('''_''') return text def A__ ( self ,A__): return " ".join(A__) def A__ ( self ,A__ ,A__ = None ,A__ = False): # Convert to TensorType if not isinstance(A__ ,A__): lowercase = TensorType(A__) # Get a function reference for the correct framework if tensor_type == TensorType.TENSORFLOW: if not is_tf_available(): raise ImportError( '''Unable to convert output to TensorFlow tensors format, TensorFlow is not installed.''') import tensorflow as tf lowercase = tf.constant lowercase = tf.is_tensor elif tensor_type == TensorType.PYTORCH: if not is_torch_available(): raise ImportError('''Unable to convert output to PyTorch tensors format, PyTorch is not installed.''') import torch lowercase = torch.tensor lowercase = torch.is_tensor elif tensor_type == TensorType.JAX: if not is_flax_available(): raise ImportError('''Unable to convert output to JAX tensors format, JAX is not installed.''') import jax.numpy as jnp # noqa: F811 lowercase = jnp.array lowercase = _is_jax else: lowercase = np.asarray lowercase = _is_numpy # Do the tensor conversion in batch try: if prepend_batch_axis: lowercase = [inputs] if not is_tensor(A__): lowercase = as_tensor(A__) except: # noqa E722 raise ValueError( '''Unable to create tensor, you should probably activate truncation and/or padding ''' '''with \'padding=True\' \'truncation=True\' to have batched tensors with the same length.''') return inputs def __call__( self ,A__ ,A__ ,A__="" ,A__="pt"): lowercase = [0, 0, 0] lowercase = [artist] * len(self.version) lowercase = [genres] * len(self.version) lowercase , lowercase , lowercase = self.tokenize(A__ ,A__ ,A__) lowercase , lowercase , lowercase = self._convert_token_to_id(A__ ,A__ ,A__) lowercase = [-INFINITY] * len(full_tokens[-1]) lowercase = [ self.convert_to_tensors( [input_ids + [artists_id[i]] + genres_ids[i] + full_tokens[i]] ,tensor_type=A__) for i in range(len(self.version)) ] return BatchEncoding({'''input_ids''': input_ids, '''attention_masks''': attention_masks}) def A__ ( self ,A__ ,A__ = None): if not os.path.isdir(A__): logger.error(f'Vocabulary path ({save_directory}) should be a directory') return lowercase = os.path.join( A__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''artists_file''']) with open(A__ ,'''w''' ,encoding='''utf-8''') as f: f.write(json.dumps(self.artists_encoder ,ensure_ascii=A__)) lowercase = os.path.join( A__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''genres_file''']) with open(A__ ,'''w''' ,encoding='''utf-8''') as f: f.write(json.dumps(self.genres_encoder ,ensure_ascii=A__)) lowercase = os.path.join( A__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''lyrics_file''']) with open(A__ ,'''w''' ,encoding='''utf-8''') as f: f.write(json.dumps(self.lyrics_encoder ,ensure_ascii=A__)) return (artists_file, genres_file, lyrics_file) def A__ ( self ,A__ ,A__ ,A__): lowercase = self.artists_decoder.get(A__) lowercase = [self.genres_decoder.get(A__) for genre in genres_index] lowercase = [self.lyrics_decoder.get(A__) for character in lyric_index] return artist, genres, lyrics
633
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowercase__ :Tuple = { "configuration_biogpt": ["BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BioGptConfig"], "tokenization_biogpt": ["BioGptTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :Union[str, Any] = [ "BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST", "BioGptForCausalLM", "BioGptForTokenClassification", "BioGptForSequenceClassification", "BioGptModel", "BioGptPreTrainedModel", ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys lowercase__ :Any = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
633
1
import copy import unittest from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_MULTIPLE_CHOICE_MAPPING, MODEL_FOR_QUESTION_ANSWERING_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, LayoutLMvaConfig, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, ) from transformers.models.layoutlmva.modeling_layoutlmva import LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class lowercase : def __init__( self ,A__ ,A__=2 ,A__=3 ,A__=4 ,A__=2 ,A__=7 ,A__=True ,A__=True ,A__=True ,A__=True ,A__=9_9 ,A__=3_6 ,A__=3 ,A__=4 ,A__=3_7 ,A__="gelu" ,A__=0.1 ,A__=0.1 ,A__=5_1_2 ,A__=1_6 ,A__=2 ,A__=0.02 ,A__=6 ,A__=6 ,A__=3 ,A__=4 ,A__=None ,A__=1_0_0_0 ,): lowercase = parent lowercase = batch_size lowercase = num_channels lowercase = image_size lowercase = patch_size lowercase = text_seq_length lowercase = is_training lowercase = use_input_mask lowercase = use_token_type_ids lowercase = use_labels lowercase = vocab_size lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_attention_heads lowercase = intermediate_size lowercase = hidden_act lowercase = hidden_dropout_prob lowercase = attention_probs_dropout_prob lowercase = max_position_embeddings lowercase = type_vocab_size lowercase = type_sequence_label_size lowercase = initializer_range lowercase = coordinate_size lowercase = shape_size lowercase = num_labels lowercase = num_choices lowercase = scope lowercase = range_bbox # LayoutLMv3's sequence length equals the number of text tokens + number of patches + 1 (we add 1 for the CLS token) lowercase = text_seq_length lowercase = (image_size // patch_size) ** 2 + 1 lowercase = self.text_seq_length + self.image_seq_length def A__ ( self): lowercase = ids_tensor([self.batch_size, self.text_seq_length] ,self.vocab_size) lowercase = ids_tensor([self.batch_size, self.text_seq_length, 4] ,self.range_bbox) # Ensure that bbox is legal for i in range(bbox.shape[0]): for j in range(bbox.shape[1]): if bbox[i, j, 3] < bbox[i, j, 1]: lowercase = bbox[i, j, 3] lowercase = bbox[i, j, 1] lowercase = t if bbox[i, j, 2] < bbox[i, j, 0]: lowercase = bbox[i, j, 2] lowercase = bbox[i, j, 0] lowercase = t lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) lowercase = None if self.use_input_mask: lowercase = random_attention_mask([self.batch_size, self.text_seq_length]) lowercase = None if self.use_token_type_ids: lowercase = ids_tensor([self.batch_size, self.text_seq_length] ,self.type_vocab_size) lowercase = None lowercase = None if self.use_labels: lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size) lowercase = ids_tensor([self.batch_size, self.text_seq_length] ,self.num_labels) lowercase = LayoutLMvaConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_range=self.initializer_range ,coordinate_size=self.coordinate_size ,shape_size=self.shape_size ,input_size=self.image_size ,patch_size=self.patch_size ,) return config, input_ids, bbox, pixel_values, token_type_ids, input_mask, sequence_labels, token_labels def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__): lowercase = LayoutLMvaModel(config=A__) model.to(A__) model.eval() # text + image lowercase = model(A__ ,pixel_values=A__) lowercase = model( A__ ,bbox=A__ ,pixel_values=A__ ,attention_mask=A__ ,token_type_ids=A__) lowercase = model(A__ ,bbox=A__ ,pixel_values=A__ ,token_type_ids=A__) lowercase = model(A__ ,bbox=A__ ,pixel_values=A__) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size)) # text only lowercase = model(A__) self.parent.assertEqual( result.last_hidden_state.shape ,(self.batch_size, self.text_seq_length, self.hidden_size)) # image only lowercase = model(pixel_values=A__) self.parent.assertEqual( result.last_hidden_state.shape ,(self.batch_size, self.image_seq_length, self.hidden_size)) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__): lowercase = self.num_labels lowercase = LayoutLMvaForSequenceClassification(A__) model.to(A__) model.eval() lowercase = model( A__ ,bbox=A__ ,pixel_values=A__ ,attention_mask=A__ ,token_type_ids=A__ ,labels=A__ ,) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels)) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__): lowercase = self.num_labels lowercase = LayoutLMvaForTokenClassification(config=A__) model.to(A__) model.eval() lowercase = model( A__ ,bbox=A__ ,pixel_values=A__ ,attention_mask=A__ ,token_type_ids=A__ ,labels=A__ ,) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.text_seq_length, self.num_labels)) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__): lowercase = LayoutLMvaForQuestionAnswering(config=A__) model.to(A__) model.eval() lowercase = model( A__ ,bbox=A__ ,pixel_values=A__ ,attention_mask=A__ ,token_type_ids=A__ ,start_positions=A__ ,end_positions=A__ ,) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length)) def A__ ( self): lowercase = self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) = config_and_inputs lowercase = { '''input_ids''': input_ids, '''bbox''': bbox, '''pixel_values''': pixel_values, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask, } return config, inputs_dict @require_torch class lowercase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): lowercase_ : List[str] =False lowercase_ : int =False lowercase_ : Union[str, Any] =False lowercase_ : Optional[Any] =( ( LayoutLMvaModel, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaForQuestionAnswering, ) if is_torch_available() else () ) lowercase_ : Union[str, Any] =( {'''document-question-answering''': LayoutLMvaForQuestionAnswering, '''feature-extraction''': LayoutLMvaModel} if is_torch_available() else {} ) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__): # `DocumentQuestionAnsweringPipeline` is expected to work with this model, but it combines the text and visual # embedding along the sequence dimension (dim 1), which causes an error during post-processing as `p_mask` has # the sequence dimension of the text embedding only. # (see the line `embedding_output = torch.cat([embedding_output, visual_embeddings], dim=1)`) return True def A__ ( self): lowercase = LayoutLMvaModelTester(self) lowercase = ConfigTester(self ,config_class=A__ ,hidden_size=3_7) def A__ ( self ,A__ ,A__ ,A__=False): lowercase = copy.deepcopy(A__) if model_class in get_values(A__): lowercase = { k: v.unsqueeze(1).expand(-1 ,self.model_tester.num_choices ,-1).contiguous() if isinstance(A__ ,torch.Tensor) and v.ndim > 1 else v for k, v in inputs_dict.items() } if return_labels: if model_class in get_values(A__): lowercase = torch.ones(self.model_tester.batch_size ,dtype=torch.long ,device=A__) elif model_class in get_values(A__): lowercase = torch.zeros( self.model_tester.batch_size ,dtype=torch.long ,device=A__) lowercase = torch.zeros( self.model_tester.batch_size ,dtype=torch.long ,device=A__) elif model_class in [ *get_values(A__), ]: lowercase = torch.zeros( self.model_tester.batch_size ,dtype=torch.long ,device=A__) elif model_class in [ *get_values(A__), ]: lowercase = torch.zeros( (self.model_tester.batch_size, self.model_tester.text_seq_length) ,dtype=torch.long ,device=A__ ,) return inputs_dict def A__ ( self): self.config_tester.run_common_tests() def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowercase = type self.model_tester.create_and_check_model(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*A__) @slow def A__ ( self): for model_name in LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase = LayoutLMvaModel.from_pretrained(A__) self.assertIsNotNone(A__) def UpperCamelCase ( ): '''simple docstring''' lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch class lowercase ( unittest.TestCase ): @cached_property def A__ ( self): return LayoutLMvaImageProcessor(apply_ocr=A__) if is_vision_available() else None @slow def A__ ( self): lowercase = LayoutLMvaModel.from_pretrained('''microsoft/layoutlmv3-base''').to(A__) lowercase = self.default_image_processor lowercase = prepare_img() lowercase = image_processor(images=A__ ,return_tensors='''pt''').pixel_values.to(A__) lowercase = torch.tensor([[1, 2]]) lowercase = torch.tensor([[1, 2, 3, 4], [5, 6, 7, 8]]).unsqueeze(0) # forward pass lowercase = model( input_ids=input_ids.to(A__) ,bbox=bbox.to(A__) ,pixel_values=pixel_values.to(A__) ,) # verify the logits lowercase = torch.Size((1, 1_9_9, 7_6_8)) self.assertEqual(outputs.last_hidden_state.shape ,A__) lowercase = torch.tensor( [[-0.0529, 0.3618, 0.1632], [-0.1587, -0.1667, -0.0400], [-0.1557, -0.1671, -0.0505]]).to(A__) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] ,A__ ,atol=1E-4))
633
import logging from transformers import PretrainedConfig lowercase__ :int = logging.getLogger(__name__) lowercase__ :Dict = { "bertabs-finetuned-cnndm": "https://huggingface.co/remi/bertabs-finetuned-cnndm-extractive-abstractive-summarization/resolve/main/config.json", } class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Optional[int] ='''bertabs''' def __init__( self ,A__=3_0_5_2_2 ,A__=5_1_2 ,A__=6 ,A__=5_1_2 ,A__=8 ,A__=5_1_2 ,A__=0.2 ,A__=6 ,A__=7_6_8 ,A__=8 ,A__=2_0_4_8 ,A__=0.2 ,**A__ ,): super().__init__(**A__) lowercase = vocab_size lowercase = max_pos lowercase = enc_layers lowercase = enc_hidden_size lowercase = enc_heads lowercase = enc_ff_size lowercase = enc_dropout lowercase = dec_layers lowercase = dec_hidden_size lowercase = dec_heads lowercase = dec_ff_size lowercase = dec_dropout
633
1
def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = [] lowercase = set({'''(''', '''[''', '''{'''} ) lowercase = set({''')''', ''']''', '''}'''} ) lowercase = {'''{''': '''}''', '''[''': ''']''', '''(''': ''')'''} for i in range(len(lowerCAmelCase__ ) ): if s[i] in open_brackets: stack.append(s[i] ) elif s[i] in closed_brackets and ( len(lowerCAmelCase__ ) == 0 or (len(lowerCAmelCase__ ) > 0 and open_to_closed[stack.pop()] != s[i]) ): return False return len(lowerCAmelCase__ ) == 0 def UpperCamelCase ( ): '''simple docstring''' lowercase = input('''Enter sequence of brackets: ''' ) if is_balanced(lowerCAmelCase__ ): print(lowerCAmelCase__ , '''is balanced''' ) else: print(lowerCAmelCase__ , '''is not balanced''' ) if __name__ == "__main__": main()
633
import warnings from diffusers import StableDiffusionInpaintPipeline as StableDiffusionInpaintPipeline # noqa F401 warnings.warn( "The `inpainting.py` script is outdated. Please use directly `from diffusers import" " StableDiffusionInpaintPipeline` instead." )
633
1
import unittest from transformers import EsmConfig, is_torch_available from transformers.testing_utils import TestCasePlus, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import EsmForMaskedLM, EsmForSequenceClassification, EsmForTokenClassification, EsmModel from transformers.models.esm.modeling_esm import ( ESM_PRETRAINED_MODEL_ARCHIVE_LIST, EsmEmbeddings, create_position_ids_from_input_ids, ) class lowercase : def __init__( self ,A__ ,A__=1_3 ,A__=7 ,A__=False ,A__=True ,A__=False ,A__=True ,A__=3_3 ,A__=3_2 ,A__=5 ,A__=4 ,A__=3_7 ,A__="gelu" ,A__=0.1 ,A__=0.1 ,A__=5_1_2 ,A__=1_6 ,A__=2 ,A__=0.02 ,A__=3 ,A__=4 ,A__=None ,): lowercase = parent lowercase = batch_size lowercase = seq_length lowercase = is_training lowercase = use_input_mask lowercase = use_token_type_ids lowercase = use_labels lowercase = vocab_size lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_attention_heads lowercase = intermediate_size lowercase = hidden_act lowercase = hidden_dropout_prob lowercase = attention_probs_dropout_prob lowercase = max_position_embeddings lowercase = type_vocab_size lowercase = type_sequence_label_size lowercase = initializer_range lowercase = num_labels lowercase = num_choices lowercase = scope def A__ ( self): lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size) lowercase = None if self.use_input_mask: lowercase = random_attention_mask([self.batch_size, self.seq_length]) lowercase = None lowercase = None lowercase = None if self.use_labels: lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size) lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels) lowercase = ids_tensor([self.batch_size] ,self.num_choices) lowercase = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def A__ ( self): return EsmConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,pad_token_id=1 ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_range=self.initializer_range ,) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__): lowercase = EsmModel(config=A__) model.to(A__) model.eval() lowercase = model(A__ ,attention_mask=A__) lowercase = model(A__) lowercase = model(A__) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertEqual(result.pooler_output.shape ,(self.batch_size, self.hidden_size)) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__): lowercase = EsmForMaskedLM(config=A__) model.to(A__) model.eval() lowercase = model(A__ ,attention_mask=A__ ,labels=A__) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size)) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__): lowercase = self.num_labels lowercase = EsmForTokenClassification(config=A__) model.to(A__) model.eval() lowercase = model(A__ ,attention_mask=A__ ,labels=A__) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels)) def A__ ( self): lowercase = self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) = config_and_inputs lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class lowercase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): lowercase_ : Any =False lowercase_ : str =( ( EsmForMaskedLM, EsmModel, EsmForSequenceClassification, EsmForTokenClassification, ) if is_torch_available() else () ) lowercase_ : List[Any] =() lowercase_ : Optional[Any] =( { '''feature-extraction''': EsmModel, '''fill-mask''': EsmForMaskedLM, '''text-classification''': EsmForSequenceClassification, '''token-classification''': EsmForTokenClassification, '''zero-shot''': EsmForSequenceClassification, } if is_torch_available() else {} ) lowercase_ : int =True def A__ ( self): lowercase = EsmModelTester(self) lowercase = ConfigTester(self ,config_class=A__ ,hidden_size=3_7) def A__ ( self): self.config_tester.run_common_tests() def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowercase = type self.model_tester.create_and_check_model(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*A__) @slow def A__ ( self): for model_name in ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase = EsmModel.from_pretrained(A__) self.assertIsNotNone(A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs()[0] lowercase = EsmEmbeddings(config=A__) lowercase = torch.as_tensor([[1_2, 3_1, 1_3, model.padding_idx]]) lowercase = torch.as_tensor( [ [ 0 + model.padding_idx + 1, 1 + model.padding_idx + 1, 2 + model.padding_idx + 1, model.padding_idx, ] ]) lowercase = create_position_ids_from_input_ids(A__ ,model.padding_idx) self.assertEqual(position_ids.shape ,expected_positions.shape) self.assertTrue(torch.all(torch.eq(A__ ,A__))) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs()[0] lowercase = EsmEmbeddings(config=A__) lowercase = torch.empty(2 ,4 ,3_0) lowercase = [ 0 + embeddings.padding_idx + 1, 1 + embeddings.padding_idx + 1, 2 + embeddings.padding_idx + 1, 3 + embeddings.padding_idx + 1, ] lowercase = torch.as_tensor([expected_single_positions, expected_single_positions]) lowercase = embeddings.create_position_ids_from_inputs_embeds(A__) self.assertEqual(position_ids.shape ,expected_positions.shape) self.assertTrue(torch.all(torch.eq(A__ ,A__))) @unittest.skip('''Esm does not support embedding resizing''') def A__ ( self): pass @unittest.skip('''Esm does not support embedding resizing''') def A__ ( self): pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''') def A__ ( self): pass @require_torch class lowercase ( SCREAMING_SNAKE_CASE__ ): @slow def A__ ( self): with torch.no_grad(): lowercase = EsmForMaskedLM.from_pretrained('''facebook/esm2_t6_8M_UR50D''') model.eval() lowercase = torch.tensor([[0, 1, 2, 3, 4, 5]]) lowercase = model(A__)[0] lowercase = 3_3 lowercase = torch.Size((1, 6, vocab_size)) self.assertEqual(output.shape ,A__) lowercase = torch.tensor( [[[8.9215, -10.5898, -6.4671], [-6.3967, -13.9114, -1.1212], [-7.7812, -13.9516, -3.7406]]]) self.assertTrue(torch.allclose(output[:, :3, :3] ,A__ ,atol=1E-4)) @slow def A__ ( self): with torch.no_grad(): lowercase = EsmModel.from_pretrained('''facebook/esm2_t6_8M_UR50D''') model.eval() lowercase = torch.tensor([[0, 6, 4, 1_3, 5, 4, 1_6, 1_2, 1_1, 7, 2]]) lowercase = model(A__)[0] # compare the actual values for a slice. lowercase = torch.tensor( [[[0.1444, 0.5413, 0.3248], [0.3034, 0.0053, 0.3108], [0.3228, -0.2499, 0.3415]]]) self.assertTrue(torch.allclose(output[:, :3, :3] ,A__ ,atol=1E-4))
633
import random import unittest import numpy as np import transformers from transformers import is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax if is_flax_available(): import os import jax.numpy as jnp from jax import jit from transformers import AutoTokenizer, FlaxAutoModelForCausalLM from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model lowercase__ :Optional[Any] = "0.12" # assumed parallelism: 8 if is_torch_available(): import torch def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=None ): '''simple docstring''' if rng is None: lowercase = random.Random() lowercase = 1 for dim in shape: total_dims *= dim lowercase = [] for _ in range(lowerCAmelCase__ ): values.append(rng.randint(0 , vocab_size - 1 ) ) lowercase = np.array(lowerCAmelCase__ , dtype=jnp.intaa ).reshape(lowerCAmelCase__ ) return output def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=None ): '''simple docstring''' lowercase = ids_tensor(lowerCAmelCase__ , vocab_size=2 , rng=lowerCAmelCase__ ) # make sure that at least one token is attended to for each batch lowercase = 1 return attn_mask @require_flax class lowercase : lowercase_ : Any =None lowercase_ : List[str] =() def A__ ( self): lowercase , lowercase = self.model_tester.prepare_config_and_inputs_for_common() # cut to half length & take max batch_size 3 lowercase = 2 lowercase = inputs['''input_ids'''].shape[-1] // 2 lowercase = inputs['''input_ids'''][:max_batch_size, :sequence_length] lowercase = jnp.ones_like(A__) lowercase = attention_mask[:max_batch_size, :sequence_length] # generate max 5 tokens lowercase = input_ids.shape[-1] + 5 if config.eos_token_id is not None and config.pad_token_id is None: # hack to allow generate for models such as GPT2 as is done in `generate()` lowercase = config.eos_token_id return config, input_ids, attention_mask, max_length @is_pt_flax_cross_test def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = False lowercase = max_length lowercase = 0 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model_class.__name__[4:] # Skip the "Flax" at the beginning lowercase = getattr(A__ ,A__) lowercase = pt_model_class(A__).eval() lowercase = load_flax_weights_in_pytorch_model(A__ ,flax_model.params) lowercase = flax_model.generate(A__).sequences lowercase = pt_model.generate(torch.tensor(A__ ,dtype=torch.long)) if flax_generation_outputs.shape[-1] > pt_generation_outputs.shape[-1]: lowercase = flax_generation_outputs[:, : pt_generation_outputs.shape[-1]] self.assertListEqual(pt_generation_outputs.numpy().tolist() ,flax_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = False lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = True lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = False lowercase = max_length lowercase = 2 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = False lowercase = max_length lowercase = 2 lowercase = 2 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[0] ,input_ids.shape[0] * config.num_return_sequences) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = True lowercase = max_length lowercase = 0.8 lowercase = 1_0 lowercase = 0.3 lowercase = 1 lowercase = 8 lowercase = 9 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = max_length lowercase = 1 lowercase = 8 lowercase = 9 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() lowercase = max_length lowercase = 2 lowercase = 1 lowercase = 8 lowercase = 9 for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() # pad attention mask on the left lowercase = attention_mask.at[(0, 0)].set(0) lowercase = False lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__ ,attention_mask=A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__ ,attention_mask=A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() # pad attention mask on the left lowercase = attention_mask.at[(0, 0)].set(0) lowercase = True lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__ ,attention_mask=A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__ ,attention_mask=A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) def A__ ( self): lowercase , lowercase , lowercase , lowercase = self._get_input_ids_and_config() # pad attention mask on the left lowercase = attention_mask.at[(0, 0)].set(0) lowercase = 2 lowercase = max_length for model_class in self.all_generative_model_classes: lowercase = model_class(A__) lowercase = model.generate(A__ ,attention_mask=A__).sequences self.assertEqual(generation_outputs.shape[-1] ,A__) lowercase = jit(model.generate) lowercase = jit_generate(A__ ,attention_mask=A__).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist()) @require_flax class lowercase ( unittest.TestCase ): def A__ ( self): lowercase = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-bert''') lowercase = FlaxAutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-bert-flax-only''') lowercase = '''Hello world''' lowercase = tokenizer(A__ ,return_tensors='''np''').input_ids # typos are quickly detected (the correct argument is `do_sample`) with self.assertRaisesRegex(A__ ,'''do_samples'''): model.generate(A__ ,do_samples=A__) # arbitrary arguments that will not be used anywhere are also not accepted with self.assertRaisesRegex(A__ ,'''foo'''): lowercase = {'''foo''': '''bar'''} model.generate(A__ ,**A__)
633
1
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy import tensorflow as tf from transformers import ( TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, TF_DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, TF_DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, BertConfig, DPRConfig, TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, ) class lowercase : def __init__( self ,A__ ,A__=1_3 ,A__=7 ,A__=True ,A__=True ,A__=True ,A__=True ,A__=9_9 ,A__=3_2 ,A__=2 ,A__=4 ,A__=3_7 ,A__="gelu" ,A__=0.1 ,A__=0.1 ,A__=5_1_2 ,A__=1_6 ,A__=2 ,A__=0.02 ,A__=3 ,A__=4 ,A__=None ,A__=0 ,): lowercase = parent lowercase = batch_size lowercase = seq_length lowercase = is_training lowercase = use_input_mask lowercase = use_token_type_ids lowercase = use_labels lowercase = vocab_size lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_attention_heads lowercase = intermediate_size lowercase = hidden_act lowercase = hidden_dropout_prob lowercase = attention_probs_dropout_prob lowercase = max_position_embeddings lowercase = type_vocab_size lowercase = type_sequence_label_size lowercase = initializer_range lowercase = num_labels lowercase = num_choices lowercase = scope lowercase = projection_dim def A__ ( self): lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size) lowercase = None if self.use_input_mask: # follow test_modeling_tf_ctrl.py lowercase = random_attention_mask([self.batch_size, self.seq_length]) lowercase = None if self.use_token_type_ids: lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size) lowercase = None lowercase = None lowercase = None if self.use_labels: lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size) lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels) lowercase = ids_tensor([self.batch_size] ,self.num_choices) lowercase = BertConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,is_decoder=A__ ,initializer_range=self.initializer_range ,) lowercase = DPRConfig(projection_dim=self.projection_dim ,**config.to_dict()) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__): lowercase = TFDPRContextEncoder(config=A__) lowercase = model(A__ ,attention_mask=A__ ,token_type_ids=A__) lowercase = model(A__ ,token_type_ids=A__) lowercase = model(A__) self.parent.assertEqual(result.pooler_output.shape ,(self.batch_size, self.projection_dim or self.hidden_size)) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__): lowercase = TFDPRQuestionEncoder(config=A__) lowercase = model(A__ ,attention_mask=A__ ,token_type_ids=A__) lowercase = model(A__ ,token_type_ids=A__) lowercase = model(A__) self.parent.assertEqual(result.pooler_output.shape ,(self.batch_size, self.projection_dim or self.hidden_size)) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__): lowercase = TFDPRReader(config=A__) lowercase = model(A__ ,attention_mask=A__) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length)) self.parent.assertEqual(result.relevance_logits.shape ,(self.batch_size,)) def A__ ( self): lowercase = self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) = config_and_inputs lowercase = {'''input_ids''': input_ids} return config, inputs_dict @require_tf class lowercase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): lowercase_ : Optional[Any] =( ( TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, ) if is_tf_available() else () ) lowercase_ : List[Any] ={'''feature-extraction''': TFDPRQuestionEncoder} if is_tf_available() else {} lowercase_ : List[Any] =False lowercase_ : int =False lowercase_ : Union[str, Any] =False lowercase_ : int =False lowercase_ : List[Any] =False def A__ ( self): lowercase = TFDPRModelTester(self) lowercase = ConfigTester(self ,config_class=A__ ,hidden_size=3_7) def A__ ( self): self.config_tester.run_common_tests() def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_context_encoder(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_question_encoder(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_reader(*A__) @slow def A__ ( self): for model_name in TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase = TFDPRContextEncoder.from_pretrained(A__) self.assertIsNotNone(A__) for model_name in TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase = TFDPRContextEncoder.from_pretrained(A__) self.assertIsNotNone(A__) for model_name in TF_DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase = TFDPRQuestionEncoder.from_pretrained(A__) self.assertIsNotNone(A__) for model_name in TF_DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase = TFDPRReader.from_pretrained(A__) self.assertIsNotNone(A__) @require_tf class lowercase ( unittest.TestCase ): @slow def A__ ( self): lowercase = TFDPRQuestionEncoder.from_pretrained('''facebook/dpr-question_encoder-single-nq-base''') lowercase = tf.constant( [[1_0_1, 7_5_9_2, 1_0_1_0, 2_0_0_3, 2_0_2_6, 3_8_9_9, 1_0_1_4_0, 1_0_2_9, 1_0_2]]) # [CLS] hello, is my dog cute? [SEP] lowercase = model(A__)[0] # embedding shape = (1, 768) # compare the actual values for a slice. lowercase = tf.constant( [ [ 0.03236253, 0.12753335, 0.16818509, 0.00279786, 0.3896933, 0.24264945, 0.2178971, -0.02335227, -0.08481959, -0.14324117, ] ]) self.assertTrue(numpy.allclose(output[:, :1_0].numpy() ,expected_slice.numpy() ,atol=1E-4))
633
import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : List[str] =(UniPCMultistepScheduler,) lowercase_ : Tuple =(('''num_inference_steps''', 25),) def A__ ( self ,**A__): lowercase = { '''num_train_timesteps''': 1_0_0_0, '''beta_start''': 0.0001, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', '''solver_order''': 2, '''solver_type''': '''bh2''', } config.update(**A__) return config def A__ ( self ,A__=0 ,**A__): lowercase = dict(self.forward_default_kwargs) lowercase = kwargs.pop('''num_inference_steps''' ,A__) lowercase = self.dummy_sample lowercase = 0.1 * sample lowercase = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) scheduler.set_timesteps(A__) # copy over dummy past residuals lowercase = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A__) lowercase = scheduler_class.from_pretrained(A__) new_scheduler.set_timesteps(A__) # copy over dummy past residuals lowercase = dummy_past_residuals[: new_scheduler.config.solver_order] lowercase , lowercase = sample, sample for t in range(A__ ,time_step + scheduler.config.solver_order + 1): lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = new_scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1E-5, "Scheduler outputs are not identical" def A__ ( self ,A__=0 ,**A__): lowercase = dict(self.forward_default_kwargs) lowercase = kwargs.pop('''num_inference_steps''' ,A__) lowercase = self.dummy_sample lowercase = 0.1 * sample lowercase = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config() lowercase = scheduler_class(**A__) scheduler.set_timesteps(A__) # copy over dummy past residuals (must be after setting timesteps) lowercase = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A__) lowercase = scheduler_class.from_pretrained(A__) # copy over dummy past residuals new_scheduler.set_timesteps(A__) # copy over dummy past residual (must be after setting timesteps) lowercase = dummy_past_residuals[: new_scheduler.config.solver_order] lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = new_scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1E-5, "Scheduler outputs are not identical" def A__ ( self ,A__=None ,**A__): if scheduler is None: lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) lowercase = 1_0 lowercase = self.dummy_model() lowercase = self.dummy_sample_deter scheduler.set_timesteps(A__) for i, t in enumerate(scheduler.timesteps): lowercase = model(A__ ,A__) lowercase = scheduler.step(A__ ,A__ ,A__).prev_sample return sample def A__ ( self): lowercase = dict(self.forward_default_kwargs) lowercase = kwargs.pop('''num_inference_steps''' ,A__) for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config() lowercase = scheduler_class(**A__) lowercase = self.dummy_sample lowercase = 0.1 * sample if num_inference_steps is not None and hasattr(A__ ,'''set_timesteps'''): scheduler.set_timesteps(A__) elif num_inference_steps is not None and not hasattr(A__ ,'''set_timesteps'''): lowercase = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) lowercase = [residual + 0.2, residual + 0.15, residual + 0.10] lowercase = dummy_past_residuals[: scheduler.config.solver_order] lowercase = scheduler.timesteps[5] lowercase = scheduler.timesteps[6] lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample lowercase = scheduler.step(A__ ,A__ ,A__ ,**A__).prev_sample self.assertEqual(output_a.shape ,sample.shape) self.assertEqual(output_a.shape ,output_a.shape) def A__ ( self): # make sure that iterating over schedulers with same config names gives same results # for defaults lowercase = UniPCMultistepScheduler(**self.get_scheduler_config()) lowercase = self.full_loop(scheduler=A__) lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.2464) < 1E-3 lowercase = DPMSolverSinglestepScheduler.from_config(scheduler.config) lowercase = DEISMultistepScheduler.from_config(scheduler.config) lowercase = DPMSolverMultistepScheduler.from_config(scheduler.config) lowercase = UniPCMultistepScheduler.from_config(scheduler.config) lowercase = self.full_loop(scheduler=A__) lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.2464) < 1E-3 def A__ ( self): for timesteps in [2_5, 5_0, 1_0_0, 9_9_9, 1_0_0_0]: self.check_over_configs(num_train_timesteps=A__) def A__ ( self): self.check_over_configs(thresholding=A__) for order in [1, 2, 3]: for solver_type in ["bh1", "bh2"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=A__ ,prediction_type=A__ ,sample_max_value=A__ ,solver_order=A__ ,solver_type=A__ ,) def A__ ( self): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=A__) def A__ ( self): for solver_type in ["bh1", "bh2"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=A__ ,solver_type=A__ ,prediction_type=A__ ,) lowercase = self.full_loop( solver_order=A__ ,solver_type=A__ ,prediction_type=A__ ,) assert not torch.isnan(A__).any(), "Samples have nan numbers" def A__ ( self): self.check_over_configs(lower_order_final=A__) self.check_over_configs(lower_order_final=A__) def A__ ( self): for num_inference_steps in [1, 2, 3, 5, 1_0, 5_0, 1_0_0, 9_9_9, 1_0_0_0]: self.check_over_forward(num_inference_steps=A__ ,time_step=0) def A__ ( self): lowercase = self.full_loop() lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.2464) < 1E-3 def A__ ( self): lowercase = self.full_loop(prediction_type='''v_prediction''') lowercase = torch.mean(torch.abs(A__)) assert abs(result_mean.item() - 0.1014) < 1E-3 def A__ ( self): lowercase = self.scheduler_classes[0] lowercase = self.get_scheduler_config(thresholding=A__ ,dynamic_thresholding_ratio=0) lowercase = scheduler_class(**A__) lowercase = 1_0 lowercase = self.dummy_model() lowercase = self.dummy_sample_deter.half() scheduler.set_timesteps(A__) for i, t in enumerate(scheduler.timesteps): lowercase = model(A__ ,A__) lowercase = scheduler.step(A__ ,A__ ,A__).prev_sample assert sample.dtype == torch.floataa def A__ ( self ,**A__): for scheduler_class in self.scheduler_classes: lowercase = self.get_scheduler_config(**A__) lowercase = scheduler_class(**A__) scheduler.set_timesteps(scheduler.config.num_train_timesteps) assert len(scheduler.timesteps.unique()) == scheduler.num_inference_steps
633
1
import argparse import os import numpy as np import tensorflow as tf import torch from transformers import BertModel def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = ('''dense.weight''', '''attention.self.query''', '''attention.self.key''', '''attention.self.value''') lowercase = ( ('''layer.''', '''layer_'''), ('''word_embeddings.weight''', '''word_embeddings'''), ('''position_embeddings.weight''', '''position_embeddings'''), ('''token_type_embeddings.weight''', '''token_type_embeddings'''), ('''.''', '''/'''), ('''LayerNorm/weight''', '''LayerNorm/gamma'''), ('''LayerNorm/bias''', '''LayerNorm/beta'''), ('''weight''', '''kernel'''), ) if not os.path.isdir(lowerCAmelCase__ ): os.makedirs(lowerCAmelCase__ ) lowercase = model.state_dict() def to_tf_var_name(lowerCAmelCase__ ): for patt, repl in iter(lowerCAmelCase__ ): lowercase = name.replace(lowerCAmelCase__ , lowerCAmelCase__ ) return f'bert/{name}' def create_tf_var(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): lowercase = tf.dtypes.as_dtype(tensor.dtype ) lowercase = tf.get_variable(dtype=lowerCAmelCase__ , shape=tensor.shape , name=lowerCAmelCase__ , initializer=tf.zeros_initializer() ) session.run(tf.variables_initializer([tf_var] ) ) session.run(lowerCAmelCase__ ) return tf_var tf.reset_default_graph() with tf.Session() as session: for var_name in state_dict: lowercase = to_tf_var_name(lowerCAmelCase__ ) lowercase = state_dict[var_name].numpy() if any(x in var_name for x in tensors_to_transpose ): lowercase = torch_tensor.T lowercase = create_tf_var(tensor=lowerCAmelCase__ , name=lowerCAmelCase__ , session=lowerCAmelCase__ ) tf.keras.backend.set_value(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase = session.run(lowerCAmelCase__ ) print(f'Successfully created {tf_name}: {np.allclose(lowerCAmelCase__ , lowerCAmelCase__ )}' ) lowercase = tf.train.Saver(tf.trainable_variables() ) saver.save(lowerCAmelCase__ , os.path.join(lowerCAmelCase__ , model_name.replace('''-''' , '''_''' ) + '''.ckpt''' ) ) def UpperCamelCase ( lowerCAmelCase__=None ): '''simple docstring''' lowercase = argparse.ArgumentParser() parser.add_argument('''--model_name''' , type=lowerCAmelCase__ , required=lowerCAmelCase__ , help='''model name e.g. bert-base-uncased''' ) parser.add_argument( '''--cache_dir''' , type=lowerCAmelCase__ , default=lowerCAmelCase__ , required=lowerCAmelCase__ , help='''Directory containing pytorch model''' ) parser.add_argument('''--pytorch_model_path''' , type=lowerCAmelCase__ , required=lowerCAmelCase__ , help='''/path/to/<pytorch-model-name>.bin''' ) parser.add_argument('''--tf_cache_dir''' , type=lowerCAmelCase__ , required=lowerCAmelCase__ , help='''Directory in which to save tensorflow model''' ) lowercase = parser.parse_args(lowerCAmelCase__ ) lowercase = BertModel.from_pretrained( pretrained_model_name_or_path=args.model_name , state_dict=torch.load(args.pytorch_model_path ) , cache_dir=args.cache_dir , ) convert_pytorch_checkpoint_to_tf(model=lowerCAmelCase__ , ckpt_dir=args.tf_cache_dir , model_name=args.model_name ) if __name__ == "__main__": main()
633
import argparse from collections import defaultdict import yaml lowercase__ :Optional[int] = "docs/source/en/_toctree.yml" def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = defaultdict(lowerCAmelCase__ ) for doc in model_doc: counts[doc["local"]] += 1 lowercase = [key for key, value in counts.items() if value > 1] lowercase = [] for duplicate_key in duplicates: lowercase = list({doc['''title'''] for doc in model_doc if doc['''local'''] == duplicate_key} ) if len(lowerCAmelCase__ ) > 1: raise ValueError( f'{duplicate_key} is present several times in the documentation table of content at ' '''`docs/source/en/_toctree.yml` with different *Title* values. Choose one of those and remove the ''' '''others.''' ) # Only add this once new_doc.append({'''local''': duplicate_key, '''title''': titles[0]} ) # Add none duplicate-keys new_doc.extend([doc for doc in model_doc if counts[doc['''local''']] == 1] ) # Sort return sorted(lowerCAmelCase__ , key=lambda lowerCAmelCase__ : s["title"].lower() ) def UpperCamelCase ( lowerCAmelCase__=False ): '''simple docstring''' with open(lowerCAmelCase__ , encoding='''utf-8''' ) as f: lowercase = yaml.safe_load(f.read() ) # Get to the API doc lowercase = 0 while content[api_idx]["title"] != "API": api_idx += 1 lowercase = content[api_idx]['''sections'''] # Then to the model doc lowercase = 0 while api_doc[model_idx]["title"] != "Models": model_idx += 1 lowercase = api_doc[model_idx]['''sections'''] lowercase = [(idx, section) for idx, section in enumerate(lowerCAmelCase__ ) if '''sections''' in section] lowercase = False for idx, modality_doc in modalities_docs: lowercase = modality_doc['''sections'''] lowercase = clean_model_doc_toc(lowerCAmelCase__ ) if old_modality_doc != new_modality_doc: lowercase = True if overwrite: lowercase = new_modality_doc if diff: if overwrite: lowercase = model_doc lowercase = api_doc with open(lowerCAmelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(yaml.dump(lowerCAmelCase__ , allow_unicode=lowerCAmelCase__ ) ) else: raise ValueError( '''The model doc part of the table of content is not properly sorted, run `make style` to fix this.''' ) if __name__ == "__main__": lowercase__ :Any = argparse.ArgumentParser() parser.add_argument("--fix_and_overwrite", action="store_true", help="Whether to fix inconsistencies.") lowercase__ :int = parser.parse_args() check_model_doc(args.fix_and_overwrite)
633
1
import argparse import math import os from copy import deepcopy import torch from audio_diffusion.models import DiffusionAttnUnetaD from diffusion import sampling from torch import nn from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel lowercase__ :Optional[int] = { "gwf-440k": { "url": "https://model-server.zqevans2.workers.dev/gwf-440k.ckpt", "sample_rate": 4_8000, "sample_size": 6_5536, }, "jmann-small-190k": { "url": "https://model-server.zqevans2.workers.dev/jmann-small-190k.ckpt", "sample_rate": 4_8000, "sample_size": 6_5536, }, "jmann-large-580k": { "url": "https://model-server.zqevans2.workers.dev/jmann-large-580k.ckpt", "sample_rate": 4_8000, "sample_size": 13_1072, }, "maestro-uncond-150k": { "url": "https://model-server.zqevans2.workers.dev/maestro-uncond-150k.ckpt", "sample_rate": 1_6000, "sample_size": 6_5536, }, "unlocked-uncond-250k": { "url": "https://model-server.zqevans2.workers.dev/unlocked-uncond-250k.ckpt", "sample_rate": 1_6000, "sample_size": 6_5536, }, "honk-140k": { "url": "https://model-server.zqevans2.workers.dev/honk-140k.ckpt", "sample_rate": 1_6000, "sample_size": 6_5536, }, } def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' return torch.atana(lowerCAmelCase__ , lowerCAmelCase__ ) / math.pi * 2 def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = torch.sin(t * math.pi / 2 ) ** 2 lowercase = (1 - sigma**2) ** 0.5 return alpha_sigma_to_t(lowerCAmelCase__ , lowerCAmelCase__ ) class lowercase ( SCREAMING_SNAKE_CASE__ ): pass class lowercase ( nn.Module ): def __init__( self ,A__): super().__init__() lowercase = DiffusionAttnUnetaD(A__ ,n_attn_layers=4) lowercase = deepcopy(self.diffusion) lowercase = torch.quasirandom.SobolEngine(1 ,scramble=A__) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = MODELS_MAP[model_name]['''url'''] os.system(f'wget {url} ./' ) return f'./{model_name}.ckpt' lowercase__ :Optional[Any] = { "1": "resnets.0", "2": "attentions.0", "3": "resnets.1", "4": "attentions.1", "5": "resnets.2", "6": "attentions.2", } lowercase__ :Tuple = { "8": "resnets.0", "9": "attentions.0", "10": "resnets.1", "11": "attentions.1", "12": "resnets.2", "13": "attentions.2", } lowercase__ :int = { "1": "resnets.0", "2": "attentions.0", "3": "resnets.1", "4": "attentions.1", "5": "resnets.2", "6": "attentions.2", "8": "resnets.3", "9": "attentions.3", "10": "resnets.4", "11": "attentions.4", "12": "resnets.5", "13": "attentions.5", } lowercase__ :Dict = { "0": "resnets.0", "1": "resnets.1", "2": "resnets.2", "4": "resnets.0", "5": "resnets.1", "6": "resnets.2", } lowercase__ :Optional[Any] = { "skip": "conv_skip", "main.0": "conv_1", "main.1": "group_norm_1", "main.3": "conv_2", "main.4": "group_norm_2", } lowercase__ :List[Any] = { "norm": "group_norm", "qkv_proj": ["query", "key", "value"], "out_proj": ["proj_attn"], } def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if name.startswith('''skip''' ): return name.replace('''skip''' , RES_CONV_MAP['''skip'''] ) # name has to be of format main.{digit} if not name.startswith('''main.''' ): raise ValueError(f'ResConvBlock error with {name}' ) return name.replace(name[:6] , RES_CONV_MAP[name[:6]] ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' for key, value in ATTN_MAP.items(): if name.startswith(lowerCAmelCase__ ) and not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): return name.replace(lowerCAmelCase__ , lowerCAmelCase__ ) elif name.startswith(lowerCAmelCase__ ): return [name.replace(lowerCAmelCase__ , lowerCAmelCase__ ) for v in value] raise ValueError(f'Attn error with {name}' ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=13 ): '''simple docstring''' lowercase = input_string if string.split('''.''' )[0] == "timestep_embed": return string.replace('''timestep_embed''' , '''time_proj''' ) lowercase = 0 if string.startswith('''net.3.''' ): depth += 1 lowercase = string[6:] elif string.startswith('''net.''' ): lowercase = string[4:] while string.startswith('''main.7.''' ): depth += 1 lowercase = string[7:] if string.startswith('''main.''' ): lowercase = string[5:] # mid block if string[:2].isdigit(): lowercase = string[:2] lowercase = string[2:] else: lowercase = string[0] lowercase = string[1:] if depth == max_depth: lowercase = MID_NUM_TO_LAYER[layer_num] lowercase = '''mid_block''' elif depth > 0 and int(lowerCAmelCase__ ) < 7: lowercase = DOWN_NUM_TO_LAYER[layer_num] lowercase = f'down_blocks.{depth}' elif depth > 0 and int(lowerCAmelCase__ ) > 7: lowercase = UP_NUM_TO_LAYER[layer_num] lowercase = f'up_blocks.{max_depth - depth - 1}' elif depth == 0: lowercase = DEPTH_0_TO_LAYER[layer_num] lowercase = f'up_blocks.{max_depth - 1}' if int(lowerCAmelCase__ ) > 3 else '''down_blocks.0''' if not string_left.startswith('''.''' ): raise ValueError(f'Naming error with {input_string} and string_left: {string_left}.' ) lowercase = string_left[1:] if "resnets" in new_layer: lowercase = convert_resconv_naming(lowerCAmelCase__ ) elif "attentions" in new_layer: lowercase = convert_attn_naming(lowerCAmelCase__ ) lowercase = new_string_left if not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): lowercase = prefix + '''.''' + new_layer + '''.''' + string_left else: lowercase = [prefix + '''.''' + new_layer + '''.''' + s for s in string_left] return new_string def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = {} for k, v in state_dict.items(): if k.endswith('''kernel''' ): # up- and downsample layers, don't have trainable weights continue lowercase = rename(lowerCAmelCase__ ) # check if we need to transform from Conv => Linear for attention if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): lowercase = transform_conv_attns(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) else: lowercase = v return new_state_dict def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' if len(lowerCAmelCase__ ) == 1: if len(v.shape ) == 3: # weight lowercase = v[:, :, 0] else: # bias lowercase = v else: # qkv matrices lowercase = v.shape[0] lowercase = trippled_shape // 3 for i in range(3 ): if len(v.shape ) == 3: lowercase = v[i * single_shape : (i + 1) * single_shape, :, 0] else: lowercase = v[i * single_shape : (i + 1) * single_shape] return new_state_dict def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = torch.device('''cuda''' if torch.cuda.is_available() else '''cpu''' ) lowercase = args.model_path.split('''/''' )[-1].split('''.''' )[0] if not os.path.isfile(args.model_path ): assert ( model_name == args.model_path ), f'Make sure to provide one of the official model names {MODELS_MAP.keys()}' lowercase = download(lowerCAmelCase__ ) lowercase = MODELS_MAP[model_name]['''sample_rate'''] lowercase = MODELS_MAP[model_name]['''sample_size'''] lowercase = Object() lowercase = sample_size lowercase = sample_rate lowercase = 0 lowercase = UNetaDModel(sample_size=lowerCAmelCase__ , sample_rate=lowerCAmelCase__ ) lowercase = diffusers_model.state_dict() lowercase = DiffusionUncond(lowerCAmelCase__ ) orig_model.load_state_dict(torch.load(args.model_path , map_location=lowerCAmelCase__ )['''state_dict'''] ) lowercase = orig_model.diffusion_ema.eval() lowercase = orig_model.state_dict() lowercase = rename_orig_weights(lowerCAmelCase__ ) lowercase = set(renamed_state_dict.keys() ) - set(diffusers_state_dict.keys() ) lowercase = set(diffusers_state_dict.keys() ) - set(renamed_state_dict.keys() ) assert len(lowerCAmelCase__ ) == 0, f'Problem with {renamed_minus_diffusers}' assert all(k.endswith('''kernel''' ) for k in list(lowerCAmelCase__ ) ), f'Problem with {diffusers_minus_renamed}' for key, value in renamed_state_dict.items(): assert ( diffusers_state_dict[key].squeeze().shape == value.squeeze().shape ), f'Shape for {key} doesn\'t match. Diffusers: {diffusers_state_dict[key].shape} vs. {value.shape}' if key == "time_proj.weight": lowercase = value.squeeze() lowercase = value diffusers_model.load_state_dict(lowerCAmelCase__ ) lowercase = 100 lowercase = 33 lowercase = IPNDMScheduler(num_train_timesteps=lowerCAmelCase__ ) lowercase = torch.manual_seed(lowerCAmelCase__ ) lowercase = torch.randn([1, 2, config.sample_size] , generator=lowerCAmelCase__ ).to(lowerCAmelCase__ ) lowercase = torch.linspace(1 , 0 , steps + 1 , device=lowerCAmelCase__ )[:-1] lowercase = get_crash_schedule(lowerCAmelCase__ ) lowercase = DanceDiffusionPipeline(unet=lowerCAmelCase__ , scheduler=lowerCAmelCase__ ) lowercase = torch.manual_seed(33 ) lowercase = pipe(num_inference_steps=lowerCAmelCase__ , generator=lowerCAmelCase__ ).audios lowercase = sampling.iplms_sample(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , {} ) lowercase = generated.clamp(-1 , 1 ) lowercase = (generated - audio).abs().sum() lowercase = (generated - audio).abs().max() if args.save: pipe.save_pretrained(args.checkpoint_path ) print('''Diff sum''' , lowerCAmelCase__ ) print('''Diff max''' , lowerCAmelCase__ ) assert diff_max < 1E-3, f'Diff max: {diff_max} is too much :-/' print(f'Conversion for {model_name} successful!' ) if __name__ == "__main__": lowercase__ :Tuple = argparse.ArgumentParser() parser.add_argument("--model_path", default=None, type=str, required=True, help="Path to the model to convert.") parser.add_argument( "--save", default=True, type=bool, required=False, help="Whether to save the converted model or not." ) parser.add_argument("--checkpoint_path", default=None, type=str, required=True, help="Path to the output model.") lowercase__ :Dict = parser.parse_args() main(args)
633
from __future__ import annotations from decimal import Decimal from math import * # noqa: F403 from sympy import diff def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 10**-10 ): '''simple docstring''' lowercase = a while True: lowercase = Decimal(lowerCAmelCase__ ) - ( Decimal(eval(lowerCAmelCase__ ) ) / Decimal(eval(str(diff(lowerCAmelCase__ ) ) ) ) # noqa: S307 ) # This number dictates the accuracy of the answer if abs(eval(lowerCAmelCase__ ) ) < precision: # noqa: S307 return float(lowerCAmelCase__ ) # Let's Execute if __name__ == "__main__": # Find root of trigonometric function # Find value of pi print(F'The root of sin(x) = 0 is {newton_raphson("sin(x)", 2)}') # Find root of polynomial print(F'The root of x**2 - 5*x + 2 = 0 is {newton_raphson("x**2 - 5*x + 2", 0.4)}') # Find Square Root of 5 print(F'The root of log(x) - 1 = 0 is {newton_raphson("log(x) - 1", 2)}') # Exponential Roots print(F'The root of exp(x) - 1 = 0 is {newton_raphson("exp(x) - 1", 0)}')
633
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowercase__ :Tuple = { "configuration_biogpt": ["BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BioGptConfig"], "tokenization_biogpt": ["BioGptTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :Union[str, Any] = [ "BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST", "BioGptForCausalLM", "BioGptForTokenClassification", "BioGptForSequenceClassification", "BioGptModel", "BioGptPreTrainedModel", ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys lowercase__ :Any = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
633
from pathlib import Path import numpy as np from PIL import Image def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase , lowercase = rgb[:, :, 0], rgb[:, :, 1], rgb[:, :, 2] return 0.29_89 * r + 0.58_70 * g + 0.11_40 * b def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return (gray > 127) & (gray <= 255) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = np.zeros_like(lowerCAmelCase__ ) lowercase = np.zeros( (image.shape[0] + kernel.shape[0] - 1, image.shape[1] + kernel.shape[1] - 1) ) # Copy image to padded image lowercase = image # Iterate over image & apply kernel for x in range(image.shape[1] ): for y in range(image.shape[0] ): lowercase = ( kernel * image_padded[y : y + kernel.shape[0], x : x + kernel.shape[1]] ).sum() lowercase = int(summation > 0 ) return output if __name__ == "__main__": # read original image lowercase__ :str = Path(__file__).resolve().parent / "image_data" / "lena.jpg" lowercase__ :List[str] = np.array(Image.open(lena_path)) # kernel to be applied lowercase__ :Union[str, Any] = np.array([[0, 1, 0], [1, 1, 1], [0, 1, 0]]) lowercase__ :Optional[int] = dilation(gray_to_binary(rgb_to_gray(lena)), structuring_element) # Save the output image lowercase__ :str = Image.fromarray(output).convert("RGB") pil_img.save("result_dilation.png")
633
1
from __future__ import annotations lowercase__ :str = tuple[int, int, int] lowercase__ :Optional[int] = tuple[str, str, str] # used alphabet -------------------------- # from string.ascii_uppercase lowercase__ :int = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" # -------------------------- default selection -------------------------- # rotors -------------------------- lowercase__ :Tuple = "EGZWVONAHDCLFQMSIPJBYUKXTR" lowercase__ :Dict = "FOBHMDKEXQNRAULPGSJVTYICZW" lowercase__ :int = "ZJXESIUQLHAVRMDOYGTNFWPBKC" # reflector -------------------------- lowercase__ :Tuple = { "A": "N", "N": "A", "B": "O", "O": "B", "C": "P", "P": "C", "D": "Q", "Q": "D", "E": "R", "R": "E", "F": "S", "S": "F", "G": "T", "T": "G", "H": "U", "U": "H", "I": "V", "V": "I", "J": "W", "W": "J", "K": "X", "X": "K", "L": "Y", "Y": "L", "M": "Z", "Z": "M", } # -------------------------- extra rotors -------------------------- lowercase__ :List[str] = "RMDJXFUWGISLHVTCQNKYPBEZOA" lowercase__ :str = "SGLCPQWZHKXAREONTFBVIYJUDM" lowercase__ :int = "HVSICLTYKQUBXDWAJZOMFGPREN" lowercase__ :List[str] = "RZWQHFMVDBKICJLNTUXAGYPSOE" lowercase__ :Union[str, Any] = "LFKIJODBEGAMQPXVUHYSTCZRWN" lowercase__ :Tuple = "KOAEGVDHXPQZMLFTYWJNBRCIUS" def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' # Checks if there are 3 unique rotors if (unique_rotsel := len(set(lowerCAmelCase__ ) )) < 3: lowercase = f'Please use 3 unique rotors (not {unique_rotsel})' raise Exception(lowerCAmelCase__ ) # Checks if rotor positions are valid lowercase , lowercase , lowercase = rotpos if not 0 < rotorposa <= len(lowerCAmelCase__ ): lowercase = f'First rotor position is not within range of 1..26 ({rotorposa}' raise ValueError(lowerCAmelCase__ ) if not 0 < rotorposa <= len(lowerCAmelCase__ ): lowercase = f'Second rotor position is not within range of 1..26 ({rotorposa})' raise ValueError(lowerCAmelCase__ ) if not 0 < rotorposa <= len(lowerCAmelCase__ ): lowercase = f'Third rotor position is not within range of 1..26 ({rotorposa})' raise ValueError(lowerCAmelCase__ ) # Validates string and returns dict lowercase = _plugboard(lowerCAmelCase__ ) return rotpos, rotsel, pbdict def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' # tests the input string if it # a) is type string # b) has even length (so pairs can be made) if not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): lowercase = f'Plugboard setting isn\'t type string ({type(lowerCAmelCase__ )})' raise TypeError(lowerCAmelCase__ ) elif len(lowerCAmelCase__ ) % 2 != 0: lowercase = f'Odd number of symbols ({len(lowerCAmelCase__ )})' raise Exception(lowerCAmelCase__ ) elif pbstring == "": return {} pbstring.replace(''' ''' , '''''' ) # Checks if all characters are unique lowercase = set() for i in pbstring: if i not in abc: lowercase = f'\'{i}\' not in list of symbols' raise Exception(lowerCAmelCase__ ) elif i in tmppbl: lowercase = f'Duplicate symbol ({i})' raise Exception(lowerCAmelCase__ ) else: tmppbl.add(lowerCAmelCase__ ) del tmppbl # Created the dictionary lowercase = {} for j in range(0 , len(lowerCAmelCase__ ) - 1 , 2 ): lowercase = pbstring[j + 1] lowercase = pbstring[j] return pb def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = (rotora, rotora, rotora) , lowerCAmelCase__ = "" , ): '''simple docstring''' lowercase = text.upper() lowercase , lowercase , lowercase = _validator( lowerCAmelCase__ , lowerCAmelCase__ , plugb.upper() ) lowercase , lowercase , lowercase = rotor_position lowercase , lowercase , lowercase = rotor_selection rotorposa -= 1 rotorposa -= 1 rotorposa -= 1 lowercase = [] # encryption/decryption process -------------------------- for symbol in text: if symbol in abc: # 1st plugboard -------------------------- if symbol in plugboard: lowercase = plugboard[symbol] # rotor ra -------------------------- lowercase = abc.index(lowerCAmelCase__ ) + rotorposa lowercase = rotora[index % len(lowerCAmelCase__ )] # rotor rb -------------------------- lowercase = abc.index(lowerCAmelCase__ ) + rotorposa lowercase = rotora[index % len(lowerCAmelCase__ )] # rotor rc -------------------------- lowercase = abc.index(lowerCAmelCase__ ) + rotorposa lowercase = rotora[index % len(lowerCAmelCase__ )] # reflector -------------------------- # this is the reason you don't need another machine to decipher lowercase = reflector[symbol] # 2nd rotors lowercase = abc[rotora.index(lowerCAmelCase__ ) - rotorposa] lowercase = abc[rotora.index(lowerCAmelCase__ ) - rotorposa] lowercase = abc[rotora.index(lowerCAmelCase__ ) - rotorposa] # 2nd plugboard if symbol in plugboard: lowercase = plugboard[symbol] # moves/resets rotor positions rotorposa += 1 if rotorposa >= len(lowerCAmelCase__ ): lowercase = 0 rotorposa += 1 if rotorposa >= len(lowerCAmelCase__ ): lowercase = 0 rotorposa += 1 if rotorposa >= len(lowerCAmelCase__ ): lowercase = 0 # else: # pass # Error could be also raised # raise ValueError( # 'Invalid symbol('+repr(symbol)+')') result.append(lowerCAmelCase__ ) return "".join(lowerCAmelCase__ ) if __name__ == "__main__": lowercase__ :int = "This is my Python script that emulates the Enigma machine from WWII." lowercase__ :List[str] = (1, 1, 1) lowercase__ :Any = "pictures" lowercase__ :str = (rotora, rotora, rotora) lowercase__ :Union[str, Any] = enigma(message, rotor_pos, rotor_sel, pb) print("Encrypted message:", en) print("Decrypted message:", enigma(en, rotor_pos, rotor_sel, pb))
633
import argparse import torch from torch import nn from transformers import MaMaaaConfig, MaMaaaForConditionalGeneration def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = [ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''decoder.output_projection.weight''', '''_float_tensor''', '''encoder.embed_positions._float_tensor''', '''decoder.embed_positions._float_tensor''', ] for k in ignore_keys: state_dict.pop(lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase = emb.weight.shape lowercase = nn.Linear(lowerCAmelCase__ , lowerCAmelCase__ , bias=lowerCAmelCase__ ) lowercase = emb.weight.data return lin_layer def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = torch.load(lowerCAmelCase__ , map_location='''cpu''' ) lowercase = mam_aaa['''args'''] or mam_aaa['''cfg''']['''model'''] lowercase = mam_aaa['''model'''] remove_ignore_keys_(lowerCAmelCase__ ) lowercase = state_dict['''encoder.embed_tokens.weight'''].shape[0] lowercase = MaMaaaConfig( vocab_size=lowerCAmelCase__ , max_position_embeddings=1024 , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , encoder_layerdrop=args.encoder_layerdrop , decoder_layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='''relu''' , ) lowercase = state_dict['''decoder.embed_tokens.weight'''] lowercase = MaMaaaForConditionalGeneration(lowerCAmelCase__ ) model.model.load_state_dict(lowerCAmelCase__ , strict=lowerCAmelCase__ ) lowercase = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": lowercase__ :Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument("fairseq_path", type=str, help="path to a model.pt on local filesystem.") parser.add_argument("pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") lowercase__ :Tuple = parser.parse_args() lowercase__ :int = convert_fairseq_mamaaa_checkpoint_from_disk(args.fairseq_pathß) model.save_pretrained(args.pytorch_dump_folder_path)
633
1
from numpy import exp, pi, sqrt def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ = 0.0 , lowerCAmelCase__ = 1.0 ): '''simple docstring''' return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
633
from __future__ import annotations from random import random class lowercase : def __init__( self ,A__ = None): lowercase = value lowercase = random() lowercase = None lowercase = None def __repr__( self): from pprint import pformat if self.left is None and self.right is None: return f'\'{self.value}: {self.prior:.5}\'' else: return pformat( {f'{self.value}: {self.prior:.5}': (self.left, self.right)} ,indent=1) def __str__( self): lowercase = str(self.value) + ''' ''' lowercase = str(self.left or '''''') lowercase = str(self.right or '''''') return value + left + right def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' if root is None: # None tree is split into 2 Nones return None, None elif root.value is None: return None, None else: if value < root.value: lowercase , lowercase = split(root.left , lowerCAmelCase__ ) return left, root else: lowercase , lowercase = split(root.right , lowerCAmelCase__ ) return root, right def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' if (not left) or (not right): # If one node is None, return the other return left or right elif left.prior < right.prior: lowercase = merge(left.right , lowerCAmelCase__ ) return left else: lowercase = merge(lowerCAmelCase__ , right.left ) return right def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = Node(lowerCAmelCase__ ) lowercase , lowercase = split(lowerCAmelCase__ , lowerCAmelCase__ ) return merge(merge(lowerCAmelCase__ , lowerCAmelCase__ ) , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase = split(lowerCAmelCase__ , value - 1 ) lowercase , lowercase = split(lowerCAmelCase__ , lowerCAmelCase__ ) return merge(lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if not root: # None return else: inorder(root.left ) print(root.value , end=''',''' ) inorder(root.right ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' for arg in args.split(): if arg[0] == "+": lowercase = insert(lowerCAmelCase__ , int(arg[1:] ) ) elif arg[0] == "-": lowercase = erase(lowerCAmelCase__ , int(arg[1:] ) ) else: print('''Unknown command''' ) return root def UpperCamelCase ( ): '''simple docstring''' lowercase = None print( '''enter numbers to create a tree, + value to add value into treap, ''' '''- value to erase all nodes with value. \'q\' to quit. ''' ) lowercase = input() while args != "q": lowercase = interact_treap(lowerCAmelCase__ , lowerCAmelCase__ ) print(lowerCAmelCase__ ) lowercase = input() print('''good by!''' ) if __name__ == "__main__": import doctest doctest.testmod() main()
633
1
import unittest from transformers import XLMConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMWithLMHeadModel, ) from transformers.models.xlm.modeling_xlm import XLM_PRETRAINED_MODEL_ARCHIVE_LIST class lowercase : def __init__( self ,A__ ,A__=1_3 ,A__=7 ,A__=True ,A__=True ,A__=True ,A__=True ,A__=True ,A__=False ,A__=False ,A__=False ,A__=2 ,A__=9_9 ,A__=0 ,A__=3_2 ,A__=5 ,A__=4 ,A__=0.1 ,A__=0.1 ,A__=5_1_2 ,A__=2 ,A__=0.02 ,A__=2 ,A__=4 ,A__="last" ,A__=True ,A__=None ,A__=0 ,): lowercase = parent lowercase = batch_size lowercase = seq_length lowercase = is_training lowercase = use_input_lengths lowercase = use_token_type_ids lowercase = use_labels lowercase = gelu_activation lowercase = sinusoidal_embeddings lowercase = causal lowercase = asm lowercase = n_langs lowercase = vocab_size lowercase = n_special lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_attention_heads lowercase = hidden_dropout_prob lowercase = attention_probs_dropout_prob lowercase = max_position_embeddings lowercase = type_sequence_label_size lowercase = initializer_range lowercase = num_labels lowercase = num_choices lowercase = summary_type lowercase = use_proj lowercase = scope lowercase = bos_token_id def A__ ( self): lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size) lowercase = random_attention_mask([self.batch_size, self.seq_length]) lowercase = None if self.use_input_lengths: lowercase = ( ids_tensor([self.batch_size] ,vocab_size=2) + self.seq_length - 2 ) # small variation of seq_length lowercase = None if self.use_token_type_ids: lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.n_langs) lowercase = None lowercase = None lowercase = None if self.use_labels: lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size) lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels) lowercase = ids_tensor([self.batch_size] ,2).float() lowercase = ids_tensor([self.batch_size] ,self.num_choices) lowercase = self.get_config() return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def A__ ( self): return XLMConfig( vocab_size=self.vocab_size ,n_special=self.n_special ,emb_dim=self.hidden_size ,n_layers=self.num_hidden_layers ,n_heads=self.num_attention_heads ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,gelu_activation=self.gelu_activation ,sinusoidal_embeddings=self.sinusoidal_embeddings ,asm=self.asm ,causal=self.causal ,n_langs=self.n_langs ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,summary_type=self.summary_type ,use_proj=self.use_proj ,num_labels=self.num_labels ,bos_token_id=self.bos_token_id ,) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,): lowercase = XLMModel(config=A__) model.to(A__) model.eval() lowercase = model(A__ ,lengths=A__ ,langs=A__) lowercase = model(A__ ,langs=A__) lowercase = model(A__) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size)) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,): lowercase = XLMWithLMHeadModel(A__) model.to(A__) model.eval() lowercase = model(A__ ,token_type_ids=A__ ,labels=A__) self.parent.assertEqual(result.loss.shape ,()) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size)) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,): lowercase = XLMForQuestionAnsweringSimple(A__) model.to(A__) model.eval() lowercase = model(A__) lowercase = model(A__ ,start_positions=A__ ,end_positions=A__) lowercase = outputs self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length)) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,): lowercase = XLMForQuestionAnswering(A__) model.to(A__) model.eval() lowercase = model(A__) lowercase = model( A__ ,start_positions=A__ ,end_positions=A__ ,cls_index=A__ ,is_impossible=A__ ,p_mask=A__ ,) lowercase = model( A__ ,start_positions=A__ ,end_positions=A__ ,cls_index=A__ ,is_impossible=A__ ,) ((lowercase) , ) = result_with_labels.to_tuple() lowercase = model(A__ ,start_positions=A__ ,end_positions=A__) ((lowercase) , ) = result_with_labels.to_tuple() self.parent.assertEqual(result_with_labels.loss.shape ,()) self.parent.assertEqual(result.start_top_log_probs.shape ,(self.batch_size, model.config.start_n_top)) self.parent.assertEqual(result.start_top_index.shape ,(self.batch_size, model.config.start_n_top)) self.parent.assertEqual( result.end_top_log_probs.shape ,(self.batch_size, model.config.start_n_top * model.config.end_n_top)) self.parent.assertEqual( result.end_top_index.shape ,(self.batch_size, model.config.start_n_top * model.config.end_n_top)) self.parent.assertEqual(result.cls_logits.shape ,(self.batch_size,)) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,): lowercase = XLMForSequenceClassification(A__) model.to(A__) model.eval() lowercase = model(A__) lowercase = model(A__ ,labels=A__) self.parent.assertEqual(result.loss.shape ,()) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size)) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,): lowercase = self.num_labels lowercase = XLMForTokenClassification(A__) model.to(A__) model.eval() lowercase = model(A__ ,attention_mask=A__ ,labels=A__) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels)) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,A__ ,): lowercase = self.num_choices lowercase = XLMForMultipleChoice(config=A__) model.to(A__) model.eval() lowercase = input_ids.unsqueeze(1).expand(-1 ,self.num_choices ,-1).contiguous() lowercase = token_type_ids.unsqueeze(1).expand(-1 ,self.num_choices ,-1).contiguous() lowercase = input_mask.unsqueeze(1).expand(-1 ,self.num_choices ,-1).contiguous() lowercase = model( A__ ,attention_mask=A__ ,token_type_ids=A__ ,labels=A__ ,) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices)) def A__ ( self): lowercase = self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) = config_and_inputs lowercase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''lengths''': input_lengths} return config, inputs_dict @require_torch class lowercase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): lowercase_ : Any =( ( XLMModel, XLMWithLMHeadModel, XLMForQuestionAnswering, XLMForSequenceClassification, XLMForQuestionAnsweringSimple, XLMForTokenClassification, XLMForMultipleChoice, ) if is_torch_available() else () ) lowercase_ : Optional[int] =( (XLMWithLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Check other models whether language generation is also applicable lowercase_ : Dict =( { '''feature-extraction''': XLMModel, '''fill-mask''': XLMWithLMHeadModel, '''question-answering''': XLMForQuestionAnsweringSimple, '''text-classification''': XLMForSequenceClassification, '''text-generation''': XLMWithLMHeadModel, '''token-classification''': XLMForTokenClassification, '''zero-shot''': XLMForSequenceClassification, } if is_torch_available() else {} ) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__): if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith('''Fast''') ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def A__ ( self ,A__ ,A__ ,A__=False): lowercase = super()._prepare_for_class(A__ ,A__ ,return_labels=A__) if return_labels: if model_class.__name__ == "XLMForQuestionAnswering": lowercase = torch.zeros( self.model_tester.batch_size ,dtype=torch.long ,device=A__) lowercase = torch.zeros( self.model_tester.batch_size ,dtype=torch.long ,device=A__) return inputs_dict def A__ ( self): lowercase = XLMModelTester(self) lowercase = ConfigTester(self ,config_class=A__ ,emb_dim=3_7) def A__ ( self): self.config_tester.run_common_tests() def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_model(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_lm_head(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_simple_qa(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_qa(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_sequence_classif(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_token_classif(*A__) def A__ ( self): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_for_multiple_choice(*A__) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__=False ,A__=1): self.assertIsInstance(A__ ,A__) self.assertListEqual( [isinstance(A__ ,A__) for iter_attentions in attentions] ,[True] * len(A__)) self.assertEqual(len(A__) ,(max_length - min_length) * num_beam_groups) for idx, iter_attentions in enumerate(A__): # adds PAD dummy token lowercase = min_length + idx + 1 lowercase = min_length + idx + 1 lowercase = ( batch_size * num_beam_groups, config.num_attention_heads, tgt_len, src_len, ) # check attn size self.assertListEqual( [layer_attention.shape for layer_attention in iter_attentions] ,[expected_shape] * len(A__)) def A__ ( self ,A__ ,A__ ,A__ ,A__ ,A__ ,A__=False ,A__=1): self.assertIsInstance(A__ ,A__) self.assertListEqual( [isinstance(A__ ,A__) for iter_hidden_states in hidden_states] ,[True] * len(A__) ,) self.assertEqual(len(A__) ,(max_length - min_length) * num_beam_groups) for idx, iter_hidden_states in enumerate(A__): # adds PAD dummy token lowercase = min_length + idx + 1 lowercase = (batch_size * num_beam_groups, seq_len, config.hidden_size) # check hidden size self.assertListEqual( [layer_hidden_states.shape for layer_hidden_states in iter_hidden_states] ,[expected_shape] * len(A__) ,) pass @slow def A__ ( self): for model_name in XLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase = XLMModel.from_pretrained(A__) self.assertIsNotNone(A__) @require_torch class lowercase ( unittest.TestCase ): @slow def A__ ( self): lowercase = XLMWithLMHeadModel.from_pretrained('''xlm-mlm-en-2048''') model.to(A__) lowercase = torch.tensor([[1_4, 4_4_7]] ,dtype=torch.long ,device=A__) # the president lowercase = [ 1_4, 4_4_7, 1_4, 4_4_7, 1_4, 4_4_7, 1_4, 4_4_7, 1_4, 4_4_7, 1_4, 4_4_7, 1_4, 4_4_7, 1_4, 4_4_7, 1_4, 4_4_7, 1_4, 4_4_7, ] # the president the president the president the president the president the president the president the president the president the president # TODO(PVP): this and other input_ids I tried for generation give pretty bad results. Not sure why. Model might just not be made for auto-regressive inference lowercase = model.generate(A__ ,do_sample=A__) self.assertListEqual(output_ids[0].cpu().numpy().tolist() ,A__)
633
def UpperCamelCase ( lowerCAmelCase__ = 1000 ): '''simple docstring''' lowercase = -1 lowercase = 0 for a in range(1 , n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N eliminating c lowercase = (n * n - 2 * a * n) // (2 * n - 2 * a) lowercase = n - a - b if c * c == (a * a + b * b): lowercase = a * b * c if candidate >= product: lowercase = candidate return product if __name__ == "__main__": print(F'{solution() = }')
633
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ :Tuple = { "configuration_instructblip": [ "INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "InstructBlipConfig", "InstructBlipQFormerConfig", "InstructBlipVisionConfig", ], "processing_instructblip": ["InstructBlipProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :List[str] = [ "INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "InstructBlipQFormerModel", "InstructBlipPreTrainedModel", "InstructBlipForConditionalGeneration", "InstructBlipVisionModel", ] if TYPE_CHECKING: from .configuration_instructblip import ( INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, InstructBlipConfig, InstructBlipQFormerConfig, InstructBlipVisionConfig, ) from .processing_instructblip import InstructBlipProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_instructblip import ( INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST, InstructBlipForConditionalGeneration, InstructBlipPreTrainedModel, InstructBlipQFormerModel, InstructBlipVisionModel, ) else: import sys lowercase__ :List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
633
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ :Tuple = { "configuration_instructblip": [ "INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "InstructBlipConfig", "InstructBlipQFormerConfig", "InstructBlipVisionConfig", ], "processing_instructblip": ["InstructBlipProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ :List[str] = [ "INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "InstructBlipQFormerModel", "InstructBlipPreTrainedModel", "InstructBlipForConditionalGeneration", "InstructBlipVisionModel", ] if TYPE_CHECKING: from .configuration_instructblip import ( INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, InstructBlipConfig, InstructBlipQFormerConfig, InstructBlipVisionConfig, ) from .processing_instructblip import InstructBlipProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_instructblip import ( INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST, InstructBlipForConditionalGeneration, InstructBlipPreTrainedModel, InstructBlipQFormerModel, InstructBlipVisionModel, ) else: import sys lowercase__ :List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
633
1
import argparse import json import os import sys import tempfile import unittest from argparse import Namespace from dataclasses import dataclass, field from enum import Enum from pathlib import Path from typing import List, Literal, Optional import yaml from transformers import HfArgumentParser, TrainingArguments from transformers.hf_argparser import make_choice_type_function, string_to_bool # Since Python 3.10, we can use the builtin `|` operator for Union types # See PEP 604: https://peps.python.org/pep-0604 lowercase__ :List[str] = sys.version_info >= (3, 10) def UpperCamelCase ( lowerCAmelCase__=None , lowerCAmelCase__=None ): '''simple docstring''' return field(default_factory=lambda: default , metadata=lowerCAmelCase__ ) @dataclass class lowercase : lowercase_ : int lowercase_ : float lowercase_ : str lowercase_ : bool @dataclass class lowercase : lowercase_ : int =42 lowercase_ : str =field(default='''toto''' , metadata={'''help''': '''help message'''} ) @dataclass class lowercase : lowercase_ : bool =False lowercase_ : bool =True lowercase_ : Optional[bool] =None class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : Union[str, Any] ='''titi''' lowercase_ : Union[str, Any] ='''toto''' class lowercase ( SCREAMING_SNAKE_CASE__ ): lowercase_ : str ='''titi''' lowercase_ : int ='''toto''' lowercase_ : Union[str, Any] =42 @dataclass class lowercase : lowercase_ : BasicEnum ="toto" def A__ ( self): lowercase = BasicEnum(self.foo) @dataclass class lowercase : lowercase_ : MixedTypeEnum ="toto" def A__ ( self): lowercase = MixedTypeEnum(self.foo) @dataclass class lowercase : lowercase_ : Optional[int] =None lowercase_ : Optional[float] =field(default=SCREAMING_SNAKE_CASE__ , metadata={'''help''': '''help message'''} ) lowercase_ : Optional[str] =None lowercase_ : Optional[List[str]] =list_field(default=[] ) lowercase_ : Optional[List[int]] =list_field(default=[] ) @dataclass class lowercase : lowercase_ : List[int] =list_field(default=[] ) lowercase_ : List[int] =list_field(default=[1, 2, 3] ) lowercase_ : List[str] =list_field(default=['''Hallo''', '''Bonjour''', '''Hello'''] ) lowercase_ : List[float] =list_field(default=[0.1, 0.2, 0.3] ) @dataclass class lowercase : lowercase_ : List[int] =field() lowercase_ : str =field() lowercase_ : BasicEnum =field() def A__ ( self): lowercase = BasicEnum(self.required_enum) @dataclass class lowercase : lowercase_ : int lowercase_ : "BasicEnum" =field() lowercase_ : "Optional[bool]" =None lowercase_ : "str" =field(default='''toto''' , metadata={'''help''': '''help message'''} ) lowercase_ : "List[str]" =list_field(default=['''Hallo''', '''Bonjour''', '''Hello'''] ) if is_python_no_less_than_3_10: @dataclass class lowercase : lowercase_ : bool =False lowercase_ : bool =True lowercase_ : bool | None =None @dataclass class lowercase : lowercase_ : int | None =None lowercase_ : float | None =field(default=SCREAMING_SNAKE_CASE__ , metadata={'''help''': '''help message'''} ) lowercase_ : str | None =None lowercase_ : list[str] | None =list_field(default=[] ) lowercase_ : list[int] | None =list_field(default=[] ) class lowercase ( unittest.TestCase ): def A__ ( self ,A__ ,A__): self.assertEqual(len(a._actions) ,len(b._actions)) for x, y in zip(a._actions ,b._actions): lowercase = {k: v for k, v in vars(A__).items() if k != '''container'''} lowercase = {k: v for k, v in vars(A__).items() if k != '''container'''} # Choices with mixed type have custom function as "type" # So we need to compare results directly for equality if xx.get('''choices''' ,A__) and yy.get('''choices''' ,A__): for expected_choice in yy["choices"] + xx["choices"]: self.assertEqual(xx['''type'''](A__) ,yy['''type'''](A__)) del xx["type"], yy["type"] self.assertEqual(A__ ,A__) def A__ ( self): lowercase = HfArgumentParser(A__) lowercase = argparse.ArgumentParser() expected.add_argument('''--foo''' ,type=A__ ,required=A__) expected.add_argument('''--bar''' ,type=A__ ,required=A__) expected.add_argument('''--baz''' ,type=A__ ,required=A__) expected.add_argument('''--flag''' ,type=A__ ,default=A__ ,const=A__ ,nargs='''?''') self.argparsersEqual(A__ ,A__) lowercase = ['''--foo''', '''1''', '''--baz''', '''quux''', '''--bar''', '''0.5'''] ((lowercase) , ) = parser.parse_args_into_dataclasses(A__ ,look_for_args_file=A__) self.assertFalse(example.flag) def A__ ( self): lowercase = HfArgumentParser(A__) lowercase = argparse.ArgumentParser() expected.add_argument('''--foo''' ,default=4_2 ,type=A__) expected.add_argument('''--baz''' ,default='''toto''' ,type=A__ ,help='''help message''') self.argparsersEqual(A__ ,A__) def A__ ( self): lowercase = argparse.ArgumentParser() expected.add_argument('''--foo''' ,type=A__ ,default=A__ ,const=A__ ,nargs='''?''') expected.add_argument('''--baz''' ,type=A__ ,default=A__ ,const=A__ ,nargs='''?''') # A boolean no_* argument always has to come after its "default: True" regular counter-part # and its default must be set to False expected.add_argument('''--no_baz''' ,action='''store_false''' ,default=A__ ,dest='''baz''') expected.add_argument('''--opt''' ,type=A__ ,default=A__) lowercase = [WithDefaultBoolExample] if is_python_no_less_than_3_10: dataclass_types.append(A__) for dataclass_type in dataclass_types: lowercase = HfArgumentParser(A__) self.argparsersEqual(A__ ,A__) lowercase = parser.parse_args([]) self.assertEqual(A__ ,Namespace(foo=A__ ,baz=A__ ,opt=A__)) lowercase = parser.parse_args(['''--foo''', '''--no_baz''']) self.assertEqual(A__ ,Namespace(foo=A__ ,baz=A__ ,opt=A__)) lowercase = parser.parse_args(['''--foo''', '''--baz''']) self.assertEqual(A__ ,Namespace(foo=A__ ,baz=A__ ,opt=A__)) lowercase = parser.parse_args(['''--foo''', '''True''', '''--baz''', '''True''', '''--opt''', '''True''']) self.assertEqual(A__ ,Namespace(foo=A__ ,baz=A__ ,opt=A__)) lowercase = parser.parse_args(['''--foo''', '''False''', '''--baz''', '''False''', '''--opt''', '''False''']) self.assertEqual(A__ ,Namespace(foo=A__ ,baz=A__ ,opt=A__)) def A__ ( self): lowercase = HfArgumentParser(A__) lowercase = argparse.ArgumentParser() expected.add_argument( '''--foo''' ,default='''toto''' ,choices=['''titi''', '''toto''', 4_2] ,type=make_choice_type_function(['''titi''', '''toto''', 4_2]) ,) self.argparsersEqual(A__ ,A__) lowercase = parser.parse_args([]) self.assertEqual(args.foo ,'''toto''') lowercase = parser.parse_args_into_dataclasses([])[0] self.assertEqual(enum_ex.foo ,MixedTypeEnum.toto) lowercase = parser.parse_args(['''--foo''', '''titi''']) self.assertEqual(args.foo ,'''titi''') lowercase = parser.parse_args_into_dataclasses(['''--foo''', '''titi'''])[0] self.assertEqual(enum_ex.foo ,MixedTypeEnum.titi) lowercase = parser.parse_args(['''--foo''', '''42''']) self.assertEqual(args.foo ,4_2) lowercase = parser.parse_args_into_dataclasses(['''--foo''', '''42'''])[0] self.assertEqual(enum_ex.foo ,MixedTypeEnum.fourtytwo) def A__ ( self): @dataclass class lowercase : lowercase_ : Literal["titi", "toto", 42] ="toto" lowercase = HfArgumentParser(A__) lowercase = argparse.ArgumentParser() expected.add_argument( '''--foo''' ,default='''toto''' ,choices=('''titi''', '''toto''', 4_2) ,type=make_choice_type_function(['''titi''', '''toto''', 4_2]) ,) self.argparsersEqual(A__ ,A__) lowercase = parser.parse_args([]) self.assertEqual(args.foo ,'''toto''') lowercase = parser.parse_args(['''--foo''', '''titi''']) self.assertEqual(args.foo ,'''titi''') lowercase = parser.parse_args(['''--foo''', '''42''']) self.assertEqual(args.foo ,4_2) def A__ ( self): lowercase = HfArgumentParser(A__) lowercase = argparse.ArgumentParser() expected.add_argument('''--foo_int''' ,nargs='''+''' ,default=[] ,type=A__) expected.add_argument('''--bar_int''' ,nargs='''+''' ,default=[1, 2, 3] ,type=A__) expected.add_argument('''--foo_str''' ,nargs='''+''' ,default=['''Hallo''', '''Bonjour''', '''Hello'''] ,type=A__) expected.add_argument('''--foo_float''' ,nargs='''+''' ,default=[0.1, 0.2, 0.3] ,type=A__) self.argparsersEqual(A__ ,A__) lowercase = parser.parse_args([]) self.assertEqual( A__ ,Namespace(foo_int=[] ,bar_int=[1, 2, 3] ,foo_str=['''Hallo''', '''Bonjour''', '''Hello'''] ,foo_float=[0.1, 0.2, 0.3]) ,) lowercase = parser.parse_args('''--foo_int 1 --bar_int 2 3 --foo_str a b c --foo_float 0.1 0.7'''.split()) self.assertEqual(A__ ,Namespace(foo_int=[1] ,bar_int=[2, 3] ,foo_str=['''a''', '''b''', '''c'''] ,foo_float=[0.1, 0.7])) def A__ ( self): lowercase = argparse.ArgumentParser() expected.add_argument('''--foo''' ,default=A__ ,type=A__) expected.add_argument('''--bar''' ,default=A__ ,type=A__ ,help='''help message''') expected.add_argument('''--baz''' ,default=A__ ,type=A__) expected.add_argument('''--ces''' ,nargs='''+''' ,default=[] ,type=A__) expected.add_argument('''--des''' ,nargs='''+''' ,default=[] ,type=A__) lowercase = [OptionalExample] if is_python_no_less_than_3_10: dataclass_types.append(A__) for dataclass_type in dataclass_types: lowercase = HfArgumentParser(A__) self.argparsersEqual(A__ ,A__) lowercase = parser.parse_args([]) self.assertEqual(A__ ,Namespace(foo=A__ ,bar=A__ ,baz=A__ ,ces=[] ,des=[])) lowercase = parser.parse_args('''--foo 12 --bar 3.14 --baz 42 --ces a b c --des 1 2 3'''.split()) self.assertEqual(A__ ,Namespace(foo=1_2 ,bar=3.14 ,baz='''42''' ,ces=['''a''', '''b''', '''c'''] ,des=[1, 2, 3])) def A__ ( self): lowercase = HfArgumentParser(A__) lowercase = argparse.ArgumentParser() expected.add_argument('''--required_list''' ,nargs='''+''' ,type=A__ ,required=A__) expected.add_argument('''--required_str''' ,type=A__ ,required=A__) expected.add_argument( '''--required_enum''' ,type=make_choice_type_function(['''titi''', '''toto''']) ,choices=['''titi''', '''toto'''] ,required=A__ ,) self.argparsersEqual(A__ ,A__) def A__ ( self): lowercase = HfArgumentParser(A__) lowercase = argparse.ArgumentParser() expected.add_argument('''--foo''' ,type=A__ ,required=A__) expected.add_argument( '''--required_enum''' ,type=make_choice_type_function(['''titi''', '''toto''']) ,choices=['''titi''', '''toto'''] ,required=A__ ,) expected.add_argument('''--opt''' ,type=A__ ,default=A__) expected.add_argument('''--baz''' ,default='''toto''' ,type=A__ ,help='''help message''') expected.add_argument('''--foo_str''' ,nargs='''+''' ,default=['''Hallo''', '''Bonjour''', '''Hello'''] ,type=A__) self.argparsersEqual(A__ ,A__) def A__ ( self): lowercase = HfArgumentParser(A__) lowercase = { '''foo''': 1_2, '''bar''': 3.14, '''baz''': '''42''', '''flag''': True, } lowercase = parser.parse_dict(A__)[0] lowercase = BasicExample(**A__) self.assertEqual(A__ ,A__) def A__ ( self): lowercase = HfArgumentParser(A__) lowercase = { '''foo''': 1_2, '''bar''': 3.14, '''baz''': '''42''', '''flag''': True, '''extra''': 4_2, } self.assertRaises(A__ ,parser.parse_dict ,A__ ,allow_extra_keys=A__) def A__ ( self): lowercase = HfArgumentParser(A__) lowercase = { '''foo''': 1_2, '''bar''': 3.14, '''baz''': '''42''', '''flag''': True, } with tempfile.TemporaryDirectory() as tmp_dir: lowercase = os.path.join(A__ ,'''temp_json''') os.mkdir(A__) with open(temp_local_path + '''.json''' ,'''w+''') as f: json.dump(A__ ,A__) lowercase = parser.parse_yaml_file(Path(temp_local_path + '''.json'''))[0] lowercase = BasicExample(**A__) self.assertEqual(A__ ,A__) def A__ ( self): lowercase = HfArgumentParser(A__) lowercase = { '''foo''': 1_2, '''bar''': 3.14, '''baz''': '''42''', '''flag''': True, } with tempfile.TemporaryDirectory() as tmp_dir: lowercase = os.path.join(A__ ,'''temp_yaml''') os.mkdir(A__) with open(temp_local_path + '''.yaml''' ,'''w+''') as f: yaml.dump(A__ ,A__) lowercase = parser.parse_yaml_file(Path(temp_local_path + '''.yaml'''))[0] lowercase = BasicExample(**A__) self.assertEqual(A__ ,A__) def A__ ( self): lowercase = HfArgumentParser(A__) self.assertIsNotNone(A__)
633
from numpy import exp, pi, sqrt def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ = 0.0 , lowerCAmelCase__ = 1.0 ): '''simple docstring''' return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
633
1
# Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import subprocess from packaging.version import Version, parse from accelerate.commands.config.config_args import default_config_file, load_config_from_file lowercase__ :Any = "Run commands across TPU VMs for initial setup before running `accelerate launch`." def UpperCamelCase ( lowerCAmelCase__=None ): '''simple docstring''' if subparsers is not None: lowercase = subparsers.add_parser('''tpu-config''' , description=_description ) else: lowercase = argparse.ArgumentParser('''Accelerate tpu-config command''' , description=_description ) # Core arguments lowercase = parser.add_argument_group( '''Config Arguments''' , '''Arguments that can be configured through `accelerate config`.''' ) config_args.add_argument( '''--config_file''' , type=lowerCAmelCase__ , default=lowerCAmelCase__ , help='''Path to the config file to use for accelerate.''' , ) config_args.add_argument( '''--tpu_name''' , default=lowerCAmelCase__ , help='''The name of the TPU to use. If not specified, will use the TPU specified in the config file.''' , ) config_args.add_argument( '''--tpu_zone''' , default=lowerCAmelCase__ , help='''The zone of the TPU to use. If not specified, will use the zone specified in the config file.''' , ) lowercase = parser.add_argument_group('''TPU Arguments''' , '''Arguments for options ran inside the TPU.''' ) pod_args.add_argument( '''--use_alpha''' , action='''store_true''' , help='''Whether to use `gcloud alpha` when running the TPU training script instead of `gcloud`.''' , ) pod_args.add_argument( '''--command_file''' , default=lowerCAmelCase__ , help='''The path to the file containing the commands to run on the pod on startup.''' , ) pod_args.add_argument( '''--command''' , action='''append''' , nargs='''+''' , help='''A command to run on the pod. Can be passed multiple times.''' , ) pod_args.add_argument( '''--install_accelerate''' , action='''store_true''' , help='''Whether to install accelerate on the pod. Defaults to False.''' , ) pod_args.add_argument( '''--accelerate_version''' , default='''latest''' , help='''The version of accelerate to install on the pod. If not specified, will use the latest pypi version. Specify \'dev\' to install from GitHub.''' , ) pod_args.add_argument( '''--debug''' , action='''store_true''' , help='''If set, will print the command that would be run instead of running it.''' ) if subparsers is not None: parser.set_defaults(func=lowerCAmelCase__ ) return parser def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = None # Get the default from the config file if it exists. if args.config_file is not None or os.path.isfile(lowerCAmelCase__ ): lowercase = load_config_from_file(args.config_file ) if not args.command_file and defaults.command_file is not None and not args.command: lowercase = defaults.command_file if not args.command and defaults.commands is not None: lowercase = defaults.commands if not args.tpu_name: lowercase = defaults.tpu_name if not args.tpu_zone: lowercase = defaults.tpu_zone if args.accelerate_version == "dev": lowercase = '''git+https://github.com/huggingface/accelerate.git''' elif args.accelerate_version == "latest": lowercase = '''accelerate -U''' elif isinstance(parse(args.accelerate_version ) , lowerCAmelCase__ ): lowercase = f'accelerate=={args.accelerate_version}' if not args.command_file and not args.command: raise ValueError('''You must specify either a command file or a command to run on the pod.''' ) if args.command_file: with open(args.command_file , '''r''' ) as f: lowercase = [f.read().splitlines()] # To turn list of lists into list of strings if isinstance(args.command[0] , lowerCAmelCase__ ): lowercase = [line for cmd in args.command for line in cmd] # Default to the shared folder and install accelerate lowercase = ['''cd /usr/share'''] if args.install_accelerate: new_cmd += [f'pip install {args.accelerate_version}'] new_cmd += args.command lowercase = '''; '''.join(lowerCAmelCase__ ) # Then send it to gcloud # Eventually try to use google-api-core to do this instead of subprocess lowercase = ['''gcloud'''] if args.use_alpha: cmd += ["alpha"] cmd += [ "compute", "tpus", "tpu-vm", "ssh", args.tpu_name, "--zone", args.tpu_zone, "--command", args.command, "--worker", "all", ] if args.debug: print(f'Running {" ".join(lowerCAmelCase__ )}' ) return subprocess.run(lowerCAmelCase__ ) print('''Successfully setup pod.''' ) def UpperCamelCase ( ): '''simple docstring''' lowercase = tpu_command_parser() lowercase = parser.parse_args() tpu_command_launcher(lowerCAmelCase__ )
633
import asyncio import os import shutil import subprocess import sys import tempfile import unittest from distutils.util import strtobool from functools import partial from pathlib import Path from typing import List, Union from unittest import mock import torch from ..state import AcceleratorState, PartialState from ..utils import ( gather, is_bnb_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_mps_available, is_safetensors_available, is_tensorboard_available, is_torch_version, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=False ): '''simple docstring''' try: lowercase = os.environ[key] except KeyError: # KEY isn't set, default to `default`. lowercase = default else: # KEY is set, convert it to True or False. try: lowercase = strtobool(lowerCAmelCase__ ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(f'If set, {key} must be yes or no.' ) return _value lowercase__ :Dict = parse_flag_from_env("RUN_SLOW", default=False) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skip('''Test was skipped''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(_run_slow_tests , '''test is slow''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(not torch.cuda.is_available() , '''test requires only a CPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.cuda.is_available() , '''test requires a GPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_xpu_available() , '''test requires a XPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_mps_available() , '''test requires a `mps` backend support in `torch`''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless( is_transformers_available() and is_datasets_available() , '''test requires the Hugging Face suite''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_bnb_available() , '''test requires the bitsandbytes library''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_tpu_available() , '''test requires TPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.cuda.device_count() == 1 , '''test requires a GPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.xpu.device_count() == 1 , '''test requires a XPU''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.cuda.device_count() > 1 , '''test requires multiple GPUs''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(torch.xpu.device_count() > 1 , '''test requires multiple XPUs''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_safetensors_available() , '''test requires safetensors''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_deepspeed_available() , '''test requires DeepSpeed''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_torch_version('''>=''' , '''1.12.0''' ) , '''test requires torch version >= 1.12.0''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__=None , lowerCAmelCase__=None ): '''simple docstring''' if test_case is None: return partial(lowerCAmelCase__ , version=lowerCAmelCase__ ) return unittest.skipUnless(is_torch_version('''>=''' , lowerCAmelCase__ ) , f'test requires torch version >= {version}' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_tensorboard_available() , '''test requires Tensorboard''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_wandb_available() , '''test requires wandb''' )(lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless(is_comet_ml_available() , '''test requires comet_ml''' )(lowerCAmelCase__ ) lowercase__ :Dict = ( any([is_wandb_available(), is_tensorboard_available()]) and not is_comet_ml_available() ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return unittest.skipUnless( _atleast_one_tracker_available , '''test requires at least one tracker to be available and for `comet_ml` to not be installed''' , )(lowerCAmelCase__ ) class lowercase ( unittest.TestCase ): lowercase_ : int =True @classmethod def A__ ( cls): lowercase = tempfile.mkdtemp() @classmethod def A__ ( cls): if os.path.exists(cls.tmpdir): shutil.rmtree(cls.tmpdir) def A__ ( self): if self.clear_on_setup: for path in Path(self.tmpdir).glob('''**/*'''): if path.is_file(): path.unlink() elif path.is_dir(): shutil.rmtree(A__) class lowercase ( unittest.TestCase ): def A__ ( self): super().tearDown() # Reset the state of the AcceleratorState singleton. AcceleratorState._reset_state() PartialState._reset_state() class lowercase ( unittest.TestCase ): def A__ ( self ,A__): lowercase = mocks if isinstance(A__ ,(tuple, list)) else [mocks] for m in self.mocks: m.start() self.addCleanup(m.stop) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = AcceleratorState() lowercase = tensor[None].clone().to(state.device ) lowercase = gather(lowerCAmelCase__ ).cpu() lowercase = tensor[0].cpu() for i in range(tensors.shape[0] ): if not torch.equal(tensors[i] , lowerCAmelCase__ ): return False return True class lowercase : def __init__( self ,A__ ,A__ ,A__): lowercase = returncode lowercase = stdout lowercase = stderr async def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' while True: lowercase = await stream.readline() if line: callback(lowerCAmelCase__ ) else: break async def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=False , lowerCAmelCase__=False ): '''simple docstring''' if echo: print('''\nRunning: ''' , ''' '''.join(lowerCAmelCase__ ) ) lowercase = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=lowerCAmelCase__ , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=lowerCAmelCase__ , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) lowercase = [] lowercase = [] def tee(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__="" ): lowercase = line.decode('''utf-8''' ).rstrip() sink.append(lowerCAmelCase__ ) if not quiet: print(lowerCAmelCase__ , lowerCAmelCase__ , file=lowerCAmelCase__ ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ asyncio.create_task(_read_stream(p.stdout , lambda lowerCAmelCase__ : tee(lowerCAmelCase__ , lowerCAmelCase__ , sys.stdout , label='''stdout:''' ) ) ), asyncio.create_task(_read_stream(p.stderr , lambda lowerCAmelCase__ : tee(lowerCAmelCase__ , lowerCAmelCase__ , sys.stderr , label='''stderr:''' ) ) ), ] , timeout=lowerCAmelCase__ , ) return _RunOutput(await p.wait() , lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=180 , lowerCAmelCase__=False , lowerCAmelCase__=True ): '''simple docstring''' lowercase = asyncio.get_event_loop() lowercase = loop.run_until_complete( _stream_subprocess(lowerCAmelCase__ , env=lowerCAmelCase__ , stdin=lowerCAmelCase__ , timeout=lowerCAmelCase__ , quiet=lowerCAmelCase__ , echo=lowerCAmelCase__ ) ) lowercase = ''' '''.join(lowerCAmelCase__ ) if result.returncode > 0: lowercase = '''\n'''.join(result.stderr ) raise RuntimeError( f'\'{cmd_str}\' failed with returncode {result.returncode}\n\n' f'The combined stderr from workers follows:\n{stderr}' ) return result class lowercase ( SCREAMING_SNAKE_CASE__ ): pass def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=False ): '''simple docstring''' try: lowercase = subprocess.check_output(lowerCAmelCase__ , stderr=subprocess.STDOUT ) if return_stdout: if hasattr(lowerCAmelCase__ , '''decode''' ): lowercase = output.decode('''utf-8''' ) return output except subprocess.CalledProcessError as e: raise SubprocessCallException( f'Command `{" ".join(lowerCAmelCase__ )}` failed with the following error:\n\n{e.output.decode()}' ) from e
633
1
import math def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ = 0 , lowerCAmelCase__ = 0 ): '''simple docstring''' lowercase = end or len(lowerCAmelCase__ ) for i in range(lowerCAmelCase__ , lowerCAmelCase__ ): lowercase = i lowercase = array[i] while temp_index != start and temp_index_value < array[temp_index - 1]: lowercase = array[temp_index - 1] temp_index -= 1 lowercase = temp_index_value return array def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): # Max Heap '''simple docstring''' lowercase = index lowercase = 2 * index + 1 # Left Node lowercase = 2 * index + 2 # Right Node if left_index < heap_size and array[largest] < array[left_index]: lowercase = left_index if right_index < heap_size and array[largest] < array[right_index]: lowercase = right_index if largest != index: lowercase , lowercase = array[largest], array[index] heapify(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = len(lowerCAmelCase__ ) for i in range(n // 2 , -1 , -1 ): heapify(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) for i in range(n - 1 , 0 , -1 ): lowercase , lowercase = array[0], array[i] heapify(lowerCAmelCase__ , 0 , lowerCAmelCase__ ) return array def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' if (array[first_index] > array[middle_index]) != ( array[first_index] > array[last_index] ): return array[first_index] elif (array[middle_index] > array[first_index]) != ( array[middle_index] > array[last_index] ): return array[middle_index] else: return array[last_index] def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' lowercase = low lowercase = high while True: while array[i] < pivot: i += 1 j -= 1 while pivot < array[j]: j -= 1 if i >= j: return i lowercase , lowercase = array[j], array[i] i += 1 def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' if len(lowerCAmelCase__ ) == 0: return array lowercase = 2 * math.ceil(math.loga(len(lowerCAmelCase__ ) ) ) lowercase = 16 return intro_sort(lowerCAmelCase__ , 0 , len(lowerCAmelCase__ ) , lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): '''simple docstring''' while end - start > size_threshold: if max_depth == 0: return heap_sort(lowerCAmelCase__ ) max_depth -= 1 lowercase = median_of_a(lowerCAmelCase__ , lowerCAmelCase__ , start + ((end - start) // 2) + 1 , end - 1 ) lowercase = partition(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) intro_sort(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) lowercase = p return insertion_sort(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) if __name__ == "__main__": import doctest doctest.testmod() lowercase__ :Optional[int] = input("Enter numbers separated by a comma : ").strip() lowercase__ :List[str] = [float(item) for item in user_input.split(",")] print(sort(unsorted))
633
import math import time from typing import Dict, List, Optional from torch.utils.data import Dataset from transformers import SeqaSeqTrainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class lowercase ( SCREAMING_SNAKE_CASE__ ): def __init__( self ,*A__ ,A__=None ,A__=None ,**A__): super().__init__(*A__ ,**A__) lowercase = eval_examples lowercase = post_process_function def A__ ( self ,A__ = None ,A__=None ,A__ = None ,A__ = "eval" ,**A__ ,): lowercase = gen_kwargs.copy() lowercase = ( gen_kwargs['''max_length'''] if gen_kwargs.get('''max_length''') is not None else self.args.generation_max_length ) lowercase = ( gen_kwargs['''num_beams'''] if gen_kwargs.get('''num_beams''') is not None else self.args.generation_num_beams ) lowercase = gen_kwargs lowercase = self.eval_dataset if eval_dataset is None else eval_dataset lowercase = self.get_eval_dataloader(A__) lowercase = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. lowercase = self.compute_metrics lowercase = None lowercase = time.time() lowercase = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: lowercase = eval_loop( A__ ,description='''Evaluation''' ,prediction_loss_only=True if compute_metrics is None else None ,ignore_keys=A__ ,metric_key_prefix=A__ ,) finally: lowercase = compute_metrics lowercase = self.args.eval_batch_size * self.args.world_size if f'{metric_key_prefix}_jit_compilation_time' in output.metrics: start_time += output.metrics[f'{metric_key_prefix}_jit_compilation_time'] output.metrics.update( speed_metrics( A__ ,A__ ,num_samples=output.num_samples ,num_steps=math.ceil(output.num_samples / total_batch_size) ,)) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default lowercase = self.post_process_function(A__ ,A__ ,A__) lowercase = self.compute_metrics(A__) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys()): if not key.startswith(f'{metric_key_prefix}_'): lowercase = metrics.pop(A__) metrics.update(output.metrics) else: lowercase = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(A__) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report()) lowercase = self.callback_handler.on_evaluate(self.args ,self.state ,self.control ,A__) return metrics def A__ ( self ,A__ ,A__ ,A__=None ,A__ = "test" ,**A__): lowercase = gen_kwargs.copy() lowercase = self.get_test_dataloader(A__) # Temporarily disable metric computation, we will do it in the loop here. lowercase = self.compute_metrics lowercase = None lowercase = time.time() lowercase = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: lowercase = eval_loop( A__ ,description='''Prediction''' ,prediction_loss_only=True if compute_metrics is None else None ,ignore_keys=A__ ,metric_key_prefix=A__ ,) finally: lowercase = compute_metrics lowercase = self.args.eval_batch_size * self.args.world_size if f'{metric_key_prefix}_jit_compilation_time' in output.metrics: start_time += output.metrics[f'{metric_key_prefix}_jit_compilation_time'] output.metrics.update( speed_metrics( A__ ,A__ ,num_samples=output.num_samples ,num_steps=math.ceil(output.num_samples / total_batch_size) ,)) if self.post_process_function is None or self.compute_metrics is None: return output lowercase = self.post_process_function(A__ ,A__ ,A__ ,'''predict''') lowercase = self.compute_metrics(A__) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys()): if not key.startswith(f'{metric_key_prefix}_'): lowercase = metrics.pop(A__) metrics.update(output.metrics) return PredictionOutput(predictions=predictions.predictions ,label_ids=predictions.label_ids ,metrics=A__)
633
1