code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" import argparse import json import torch from diffusers import DDPMScheduler, LDMPipeline, UNetaDModel, VQModel def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__=1 ): '''simple docstring''' if n_shave_prefix_segments >= 0: return ".".join(path.split(""".""" )[n_shave_prefix_segments:] ) else: return ".".join(path.split(""".""" )[:n_shave_prefix_segments] ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__=0 ): '''simple docstring''' _a : Any = [] for old_item in old_list: _a : Optional[Any] = old_item.replace("""in_layers.0""" , """norm1""" ) _a : Optional[int] = new_item.replace("""in_layers.2""" , """conv1""" ) _a : Any = new_item.replace("""out_layers.0""" , """norm2""" ) _a : str = new_item.replace("""out_layers.3""" , """conv2""" ) _a : Tuple = new_item.replace("""emb_layers.1""" , """time_emb_proj""" ) _a : Dict = new_item.replace("""skip_connection""" , """conv_shortcut""" ) _a : Dict = shave_segments(UpperCamelCase__ , n_shave_prefix_segments=UpperCamelCase__ ) mapping.append({"""old""": old_item, """new""": new_item} ) return mapping def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__=0 ): '''simple docstring''' _a : Any = [] for old_item in old_list: _a : Tuple = old_item _a : Dict = new_item.replace("""norm.weight""" , """group_norm.weight""" ) _a : Tuple = new_item.replace("""norm.bias""" , """group_norm.bias""" ) _a : Optional[int] = new_item.replace("""proj_out.weight""" , """proj_attn.weight""" ) _a : List[str] = new_item.replace("""proj_out.bias""" , """proj_attn.bias""" ) _a : Any = shave_segments(UpperCamelCase__ , n_shave_prefix_segments=UpperCamelCase__ ) mapping.append({"""old""": old_item, """new""": new_item} ) return mapping def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=None ): '''simple docstring''' assert isinstance(UpperCamelCase__ , UpperCamelCase__ ), "Paths should be a list of dicts containing 'old' and 'new' keys." # Splits the attention layers into three variables. if attention_paths_to_split is not None: for path, path_map in attention_paths_to_split.items(): _a : Union[str, Any] = old_checkpoint[path] _a : int = old_tensor.shape[0] // 3 _a : Optional[int] = (-1, channels) if len(old_tensor.shape ) == 3 else (-1) _a : str = old_tensor.shape[0] // config["""num_head_channels"""] // 3 _a : int = old_tensor.reshape((num_heads, 3 * channels // num_heads) + old_tensor.shape[1:] ) _a , _a , _a : Union[str, Any] = old_tensor.split(channels // num_heads , dim=1 ) _a : List[Any] = query.reshape(UpperCamelCase__ ) _a : Tuple = key.reshape(UpperCamelCase__ ) _a : Tuple = value.reshape(UpperCamelCase__ ) for path in paths: _a : Tuple = path["""new"""] # These have already been assigned if attention_paths_to_split is not None and new_path in attention_paths_to_split: continue # Global renaming happens here _a : str = new_path.replace("""middle_block.0""" , """mid_block.resnets.0""" ) _a : Union[str, Any] = new_path.replace("""middle_block.1""" , """mid_block.attentions.0""" ) _a : int = new_path.replace("""middle_block.2""" , """mid_block.resnets.1""" ) if additional_replacements is not None: for replacement in additional_replacements: _a : Optional[int] = new_path.replace(replacement["""old"""] , replacement["""new"""] ) # proj_attn.weight has to be converted from conv 1D to linear if "proj_attn.weight" in new_path: _a : int = old_checkpoint[path["""old"""]][:, :, 0] else: _a : Optional[int] = old_checkpoint[path["""old"""]] def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : int = {} _a : int = checkpoint["""time_embed.0.weight"""] _a : Optional[Any] = checkpoint["""time_embed.0.bias"""] _a : Dict = checkpoint["""time_embed.2.weight"""] _a : List[Any] = checkpoint["""time_embed.2.bias"""] _a : Dict = checkpoint["""input_blocks.0.0.weight"""] _a : List[str] = checkpoint["""input_blocks.0.0.bias"""] _a : int = checkpoint["""out.0.weight"""] _a : str = checkpoint["""out.0.bias"""] _a : int = checkpoint["""out.2.weight"""] _a : Optional[int] = checkpoint["""out.2.bias"""] # Retrieves the keys for the input blocks only _a : int = len({""".""".join(layer.split(""".""" )[:2] ) for layer in checkpoint if """input_blocks""" in layer} ) _a : List[Any] = { layer_id: [key for key in checkpoint if F"""input_blocks.{layer_id}""" in key] for layer_id in range(UpperCamelCase__ ) } # Retrieves the keys for the middle blocks only _a : Union[str, Any] = len({""".""".join(layer.split(""".""" )[:2] ) for layer in checkpoint if """middle_block""" in layer} ) _a : Optional[int] = { layer_id: [key for key in checkpoint if F"""middle_block.{layer_id}""" in key] for layer_id in range(UpperCamelCase__ ) } # Retrieves the keys for the output blocks only _a : Optional[Any] = len({""".""".join(layer.split(""".""" )[:2] ) for layer in checkpoint if """output_blocks""" in layer} ) _a : List[str] = { layer_id: [key for key in checkpoint if F"""output_blocks.{layer_id}""" in key] for layer_id in range(UpperCamelCase__ ) } for i in range(1 , UpperCamelCase__ ): _a : int = (i - 1) // (config["""num_res_blocks"""] + 1) _a : Optional[Any] = (i - 1) % (config["""num_res_blocks"""] + 1) _a : Dict = [key for key in input_blocks[i] if F"""input_blocks.{i}.0""" in key] _a : List[str] = [key for key in input_blocks[i] if F"""input_blocks.{i}.1""" in key] if F"""input_blocks.{i}.0.op.weight""" in checkpoint: _a : Any = checkpoint[ F"""input_blocks.{i}.0.op.weight""" ] _a : int = checkpoint[ F"""input_blocks.{i}.0.op.bias""" ] continue _a : Union[str, Any] = renew_resnet_paths(UpperCamelCase__ ) _a : List[Any] = {"""old""": F"""input_blocks.{i}.0""", """new""": F"""down_blocks.{block_id}.resnets.{layer_in_block_id}"""} _a : Union[str, Any] = {"""old""": """resnets.2.op""", """new""": """downsamplers.0.op"""} assign_to_checkpoint( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , additional_replacements=[meta_path, resnet_op] , config=UpperCamelCase__ ) if len(UpperCamelCase__ ): _a : str = renew_attention_paths(UpperCamelCase__ ) _a : List[str] = { """old""": F"""input_blocks.{i}.1""", """new""": F"""down_blocks.{block_id}.attentions.{layer_in_block_id}""", } _a : Optional[Any] = { F"""input_blocks.{i}.1.qkv.bias""": { """key""": F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.key.bias""", """query""": F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.query.bias""", """value""": F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.value.bias""", }, F"""input_blocks.{i}.1.qkv.weight""": { """key""": F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.key.weight""", """query""": F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.query.weight""", """value""": F"""down_blocks.{block_id}.attentions.{layer_in_block_id}.value.weight""", }, } assign_to_checkpoint( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , additional_replacements=[meta_path] , attention_paths_to_split=UpperCamelCase__ , config=UpperCamelCase__ , ) _a : List[str] = middle_blocks[0] _a : int = middle_blocks[1] _a : Union[str, Any] = middle_blocks[2] _a : Tuple = renew_resnet_paths(UpperCamelCase__ ) assign_to_checkpoint(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , config=UpperCamelCase__ ) _a : Optional[int] = renew_resnet_paths(UpperCamelCase__ ) assign_to_checkpoint(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , config=UpperCamelCase__ ) _a : List[Any] = renew_attention_paths(UpperCamelCase__ ) _a : Dict = { """middle_block.1.qkv.bias""": { """key""": """mid_block.attentions.0.key.bias""", """query""": """mid_block.attentions.0.query.bias""", """value""": """mid_block.attentions.0.value.bias""", }, """middle_block.1.qkv.weight""": { """key""": """mid_block.attentions.0.key.weight""", """query""": """mid_block.attentions.0.query.weight""", """value""": """mid_block.attentions.0.value.weight""", }, } assign_to_checkpoint( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , attention_paths_to_split=UpperCamelCase__ , config=UpperCamelCase__ ) for i in range(UpperCamelCase__ ): _a : Any = i // (config["""num_res_blocks"""] + 1) _a : int = i % (config["""num_res_blocks"""] + 1) _a : Any = [shave_segments(UpperCamelCase__ , 2 ) for name in output_blocks[i]] _a : Any = {} for layer in output_block_layers: _a , _a : Any = layer.split(""".""" )[0], shave_segments(UpperCamelCase__ , 1 ) if layer_id in output_block_list: output_block_list[layer_id].append(UpperCamelCase__ ) else: _a : Optional[Any] = [layer_name] if len(UpperCamelCase__ ) > 1: _a : int = [key for key in output_blocks[i] if F"""output_blocks.{i}.0""" in key] _a : Union[str, Any] = [key for key in output_blocks[i] if F"""output_blocks.{i}.1""" in key] _a : Optional[Any] = renew_resnet_paths(UpperCamelCase__ ) _a : Any = renew_resnet_paths(UpperCamelCase__ ) _a : List[str] = {"""old""": F"""output_blocks.{i}.0""", """new""": F"""up_blocks.{block_id}.resnets.{layer_in_block_id}"""} assign_to_checkpoint(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , additional_replacements=[meta_path] , config=UpperCamelCase__ ) if ["conv.weight", "conv.bias"] in output_block_list.values(): _a : int = list(output_block_list.values() ).index(["""conv.weight""", """conv.bias"""] ) _a : int = checkpoint[ F"""output_blocks.{i}.{index}.conv.weight""" ] _a : Optional[int] = checkpoint[ F"""output_blocks.{i}.{index}.conv.bias""" ] # Clear attentions as they have been attributed above. if len(UpperCamelCase__ ) == 2: _a : Dict = [] if len(UpperCamelCase__ ): _a : Tuple = renew_attention_paths(UpperCamelCase__ ) _a : Tuple = { """old""": F"""output_blocks.{i}.1""", """new""": F"""up_blocks.{block_id}.attentions.{layer_in_block_id}""", } _a : Optional[int] = { F"""output_blocks.{i}.1.qkv.bias""": { """key""": F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.key.bias""", """query""": F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.query.bias""", """value""": F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.value.bias""", }, F"""output_blocks.{i}.1.qkv.weight""": { """key""": F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.key.weight""", """query""": F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.query.weight""", """value""": F"""up_blocks.{block_id}.attentions.{layer_in_block_id}.value.weight""", }, } assign_to_checkpoint( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , additional_replacements=[meta_path] , attention_paths_to_split=to_split if any("""qkv""" in key for key in attentions ) else None , config=UpperCamelCase__ , ) else: _a : Union[str, Any] = renew_resnet_paths(UpperCamelCase__ , n_shave_prefix_segments=1 ) for path in resnet_0_paths: _a : List[Any] = """.""".join(["""output_blocks""", str(UpperCamelCase__ ), path["""old"""]] ) _a : Optional[int] = """.""".join(["""up_blocks""", str(UpperCamelCase__ ), """resnets""", str(UpperCamelCase__ ), path["""new"""]] ) _a : List[str] = checkpoint[old_path] return new_checkpoint if __name__ == "__main__": _snake_case = argparse.ArgumentParser() parser.add_argument( '--checkpoint_path', default=None, type=str, required=True, help='Path to the checkpoint to convert.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help='The config json file corresponding to the architecture.', ) parser.add_argument('--dump_path', default=None, type=str, required=True, help='Path to the output model.') _snake_case = parser.parse_args() _snake_case = torch.load(args.checkpoint_path) with open(args.config_file) as f: _snake_case = json.loads(f.read()) _snake_case = convert_ldm_checkpoint(checkpoint, config) if "ldm" in config: del config["ldm"] _snake_case = UNetaDModel(**config) model.load_state_dict(converted_checkpoint) try: _snake_case = DDPMScheduler.from_config('/'.join(args.checkpoint_path.split('/')[:-1])) _snake_case = VQModel.from_pretrained('/'.join(args.checkpoint_path.split('/')[:-1])) _snake_case = LDMPipeline(unet=model, scheduler=scheduler, vae=vqvae) pipe.save_pretrained(args.dump_path) except: # noqa: E722 model.save_pretrained(args.dump_path)
294
"""simple docstring""" from __future__ import annotations def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' print(F"""Vertex\tShortest Distance from vertex {src}""" ) for i, d in enumerate(UpperCamelCase__ ): print(F"""{i}\t\t{d}""" ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' for j in range(UpperCamelCase__ ): _a , _a , _a : List[str] = (graph[j][k] for k in ["""src""", """dst""", """weight"""]) if distance[u] != float("""inf""" ) and distance[u] + w < distance[v]: return True return False def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Dict = [float("""inf""" )] * vertex_count _a : Any = 0.0 for _ in range(vertex_count - 1 ): for j in range(UpperCamelCase__ ): _a , _a , _a : List[Any] = (graph[j][k] for k in ["""src""", """dst""", """weight"""]) if distance[u] != float("""inf""" ) and distance[u] + w < distance[v]: _a : Any = distance[u] + w _a : Union[str, Any] = check_negative_cycle(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) if negative_cycle_exists: raise Exception("""Negative cycle found""" ) return distance if __name__ == "__main__": import doctest doctest.testmod() _snake_case = int(input('Enter number of vertices: ').strip()) _snake_case = int(input('Enter number of edges: ').strip()) _snake_case = [{} for _ in range(E)] for i in range(E): print('Edge ', i + 1) _snake_case , _snake_case , _snake_case = ( int(x) for x in input('Enter source, destination, weight: ').strip().split(' ') ) _snake_case = {'src': src, 'dst': dest, 'weight': weight} _snake_case = int(input('\nEnter shortest path source:').strip()) _snake_case = bellman_ford(graph, V, E, source) print_distance(shortest_distance, 0)
294
1
"""simple docstring""" import inspect import unittest import warnings from math import ceil, floor from transformers import LevitConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_MAPPING, LevitForImageClassification, LevitForImageClassificationWithTeacher, LevitModel, ) from transformers.models.levit.modeling_levit import LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class UpperCamelCase ( snake_case_ ): def _lowercase ( self : Dict ) -> int: _a : List[Any] = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(UpperCAmelCase__ , """hidden_sizes""" ) ) self.parent.assertTrue(hasattr(UpperCAmelCase__ , """num_attention_heads""" ) ) class UpperCamelCase : def __init__( self : Union[str, Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : str=13 , UpperCAmelCase__ : Optional[int]=64 , UpperCAmelCase__ : Tuple=3 , UpperCAmelCase__ : Any=3 , UpperCAmelCase__ : int=2 , UpperCAmelCase__ : Tuple=1 , UpperCAmelCase__ : Optional[Any]=16 , UpperCAmelCase__ : Tuple=[128, 256, 384] , UpperCAmelCase__ : Any=[4, 6, 8] , UpperCAmelCase__ : Dict=[2, 3, 4] , UpperCAmelCase__ : List[Any]=[16, 16, 16] , UpperCAmelCase__ : Optional[int]=0 , UpperCAmelCase__ : Union[str, Any]=[2, 2, 2] , UpperCAmelCase__ : List[str]=[2, 2, 2] , UpperCAmelCase__ : List[Any]=0.0_2 , UpperCAmelCase__ : Optional[int]=True , UpperCAmelCase__ : Union[str, Any]=True , UpperCAmelCase__ : Tuple=2 , ) -> str: _a : Tuple = parent _a : List[Any] = batch_size _a : List[Any] = image_size _a : Tuple = num_channels _a : str = kernel_size _a : List[Any] = stride _a : Union[str, Any] = padding _a : Any = hidden_sizes _a : List[Any] = num_attention_heads _a : Union[str, Any] = depths _a : List[str] = key_dim _a : Union[str, Any] = drop_path_rate _a : List[str] = patch_size _a : Optional[int] = attention_ratio _a : List[str] = mlp_ratio _a : List[str] = initializer_range _a : int = [ ["""Subsample""", key_dim[0], hidden_sizes[0] // key_dim[0], 4, 2, 2], ["""Subsample""", key_dim[0], hidden_sizes[1] // key_dim[0], 4, 2, 2], ] _a : Tuple = is_training _a : Union[str, Any] = use_labels _a : Optional[Any] = num_labels _a : str = initializer_range def _lowercase ( self : int ) -> List[str]: _a : str = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _a : List[Any] = None if self.use_labels: _a : Union[str, Any] = ids_tensor([self.batch_size] , self.num_labels ) _a : List[Any] = self.get_config() return config, pixel_values, labels def _lowercase ( self : Tuple ) -> int: return LevitConfig( image_size=self.image_size , num_channels=self.num_channels , kernel_size=self.kernel_size , stride=self.stride , padding=self.padding , patch_size=self.patch_size , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , depths=self.depths , key_dim=self.key_dim , drop_path_rate=self.drop_path_rate , mlp_ratio=self.mlp_ratio , attention_ratio=self.attention_ratio , initializer_range=self.initializer_range , down_ops=self.down_ops , ) def _lowercase ( self : Tuple , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : int , UpperCAmelCase__ : Any ) -> Union[str, Any]: _a : List[str] = LevitModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() _a : str = model(UpperCAmelCase__ ) _a : Optional[Any] = (self.image_size, self.image_size) _a , _a : Union[str, Any] = image_size[0], image_size[1] for _ in range(4 ): _a : List[str] = floor(((height + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) _a : Dict = floor(((width + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, ceil(height / 4 ) * ceil(width / 4 ), self.hidden_sizes[-1]) , ) def _lowercase ( self : Tuple , UpperCAmelCase__ : str , UpperCAmelCase__ : str , UpperCAmelCase__ : Dict ) -> int: _a : List[Any] = self.num_labels _a : Optional[Any] = LevitForImageClassification(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() _a : List[str] = model(UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowercase ( self : Tuple ) -> Union[str, Any]: _a : Tuple = self.prepare_config_and_inputs() _a , _a , _a : Union[str, Any] = config_and_inputs _a : Dict = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class UpperCamelCase ( snake_case_ , snake_case_ , unittest.TestCase ): UpperCamelCase : int = ( (LevitModel, LevitForImageClassification, LevitForImageClassificationWithTeacher) if is_torch_available() else () ) UpperCamelCase : int = ( { '''feature-extraction''': LevitModel, '''image-classification''': (LevitForImageClassification, LevitForImageClassificationWithTeacher), } if is_torch_available() else {} ) UpperCamelCase : List[str] = False UpperCamelCase : str = False UpperCamelCase : List[Any] = False UpperCamelCase : Any = False UpperCamelCase : List[Any] = False def _lowercase ( self : Optional[Any] ) -> List[str]: _a : List[Any] = LevitModelTester(self ) _a : str = ConfigTester(self , config_class=UpperCAmelCase__ , has_text_modality=UpperCAmelCase__ , hidden_size=37 ) def _lowercase ( self : int ) -> Optional[int]: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _lowercase ( self : Optional[Any] ) -> Optional[Any]: return @unittest.skip(reason="""Levit does not use inputs_embeds""" ) def _lowercase ( self : Tuple ) -> str: pass @unittest.skip(reason="""Levit does not support input and output embeddings""" ) def _lowercase ( self : Tuple ) -> int: pass @unittest.skip(reason="""Levit does not output attentions""" ) def _lowercase ( self : Optional[int] ) -> Any: pass def _lowercase ( self : Dict ) -> List[str]: _a , _a : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : List[str] = model_class(UpperCAmelCase__ ) _a : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _a : Dict = [*signature.parameters.keys()] _a : List[Any] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCAmelCase__ ) def _lowercase ( self : Optional[int] ) -> str: def check_hidden_states_output(UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : str ): _a : Union[str, Any] = model_class(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() with torch.no_grad(): _a : Dict = model(**self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ ) ) _a : Any = outputs.hidden_states _a : Any = len(self.model_tester.depths ) + 1 self.assertEqual(len(UpperCAmelCase__ ) , UpperCAmelCase__ ) _a : str = (self.model_tester.image_size, self.model_tester.image_size) _a , _a : List[str] = image_size[0], image_size[1] for _ in range(4 ): _a : List[str] = floor( ( (height + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) _a : List[Any] = floor( ( (width + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [ height * width, self.model_tester.hidden_sizes[0], ] , ) _a , _a : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Optional[int] = True check_hidden_states_output(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _a : str = True check_hidden_states_output(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def _lowercase ( self : int ) -> List[str]: pass def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[Any]=False ) -> List[str]: _a : Any = super()._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ , return_labels=UpperCAmelCase__ ) if return_labels: if model_class.__name__ == "LevitForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def _lowercase ( self : Dict ) -> Any: _a : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase__ ) def _lowercase ( self : Optional[int] ) -> Tuple: _a : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCAmelCase__ ) def _lowercase ( self : Any ) -> Optional[Any]: if not self.model_tester.is_training: return _a , _a : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() _a : Union[str, Any] = True for model_class in self.all_model_classes: # LevitForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(UpperCAmelCase__ ) or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue _a : str = model_class(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.train() _a : Tuple = self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ , return_labels=UpperCAmelCase__ ) _a : int = model(**UpperCAmelCase__ ).loss loss.backward() def _lowercase ( self : Any ) -> List[Any]: _a , _a : Dict = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return _a : List[Any] = False _a : str = True for model_class in self.all_model_classes: if model_class in get_values(UpperCAmelCase__ ) or not model_class.supports_gradient_checkpointing: continue # LevitForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "LevitForImageClassificationWithTeacher": continue _a : str = model_class(UpperCAmelCase__ ) model.gradient_checkpointing_enable() model.to(UpperCAmelCase__ ) model.train() _a : Union[str, Any] = self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ , return_labels=UpperCAmelCase__ ) _a : Any = model(**UpperCAmelCase__ ).loss loss.backward() def _lowercase ( self : Optional[Any] ) -> Tuple: _a , _a : str = self.model_tester.prepare_config_and_inputs_for_common() _a : Any = [ {"""title""": """multi_label_classification""", """num_labels""": 2, """dtype""": torch.float}, {"""title""": """single_label_classification""", """num_labels""": 1, """dtype""": torch.long}, {"""title""": """regression""", """num_labels""": 1, """dtype""": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(UpperCAmelCase__ ), ] or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=f"""Testing {model_class} with {problem_type['title']}""" ): _a : Optional[int] = problem_type["""title"""] _a : str = problem_type["""num_labels"""] _a : List[str] = model_class(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.train() _a : List[str] = self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ , return_labels=UpperCAmelCase__ ) if problem_type["num_labels"] > 1: _a : str = inputs["""labels"""].unsqueeze(1 ).repeat(1 , problem_type["""num_labels"""] ) _a : Optional[Any] = inputs["""labels"""].to(problem_type["""dtype"""] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=UpperCAmelCase__ ) as warning_list: _a : Optional[int] = model(**UpperCAmelCase__ ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( f"""Something is going wrong in the regression problem: intercepted {w.message}""" ) loss.backward() @slow def _lowercase ( self : Optional[Any] ) -> int: for model_name in LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _a : Optional[int] = LevitModel.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) def lowerCAmelCase__ ( ): '''simple docstring''' _a : Tuple = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class UpperCamelCase ( unittest.TestCase ): @cached_property def _lowercase ( self : List[str] ) -> List[str]: return LevitImageProcessor.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def _lowercase ( self : Any ) -> str: _a : List[str] = LevitForImageClassificationWithTeacher.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to( UpperCAmelCase__ ) _a : Optional[int] = self.default_image_processor _a : int = prepare_img() _a : Dict = image_processor(images=UpperCAmelCase__ , return_tensors="""pt""" ).to(UpperCAmelCase__ ) # forward pass with torch.no_grad(): _a : Union[str, Any] = model(**UpperCAmelCase__ ) # verify the logits _a : int = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase__ ) _a : List[Any] = torch.tensor([1.0_4_4_8, -0.3_7_4_5, -1.8_3_1_7] ).to(UpperCAmelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCAmelCase__ , atol=1E-4 ) )
294
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _snake_case = { 'configuration_transfo_xl': ['TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TransfoXLConfig'], 'tokenization_transfo_xl': ['TransfoXLCorpus', 'TransfoXLTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST', 'AdaptiveEmbedding', 'TransfoXLForSequenceClassification', 'TransfoXLLMHeadModel', 'TransfoXLModel', 'TransfoXLPreTrainedModel', 'load_tf_weights_in_transfo_xl', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFAdaptiveEmbedding', 'TFTransfoXLForSequenceClassification', 'TFTransfoXLLMHeadModel', 'TFTransfoXLMainLayer', 'TFTransfoXLModel', 'TFTransfoXLPreTrainedModel', ] if TYPE_CHECKING: from .configuration_transfo_xl import TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, TransfoXLConfig from .tokenization_transfo_xl import TransfoXLCorpus, TransfoXLTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_transfo_xl import ( TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, AdaptiveEmbedding, TransfoXLForSequenceClassification, TransfoXLLMHeadModel, TransfoXLModel, TransfoXLPreTrainedModel, load_tf_weights_in_transfo_xl, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_transfo_xl import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFAdaptiveEmbedding, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLMainLayer, TFTransfoXLModel, TFTransfoXLPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
294
1
"""simple docstring""" def lowerCAmelCase__ ( UpperCamelCase__=2_8_1_2_3 ): '''simple docstring''' _a : int = [1] * (limit + 1) for i in range(2 , int(limit**0.5 ) + 1 ): sum_divs[i * i] += i for k in range(i + 1 , limit // i + 1 ): sum_divs[k * i] += k + i _a : Optional[int] = set() _a : Union[str, Any] = 0 for n in range(1 , limit + 1 ): if sum_divs[n] > n: abundants.add(UpperCamelCase__ ) if not any((n - a in abundants) for a in abundants ): res += n return res if __name__ == "__main__": print(solution())
294
"""simple docstring""" def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , ): '''simple docstring''' _a : Optional[Any] = [redshift, radiation_density, matter_density, dark_energy] if any(p < 0 for p in parameters ): raise ValueError("""All input parameters must be positive""" ) if any(p > 1 for p in parameters[1:4] ): raise ValueError("""Relative densities cannot be greater than one""" ) else: _a : Tuple = 1 - (matter_density + radiation_density + dark_energy) _a : int = ( radiation_density * (redshift + 1) ** 4 + matter_density * (redshift + 1) ** 3 + curvature * (redshift + 1) ** 2 + dark_energy ) _a : List[str] = hubble_constant * e_a ** (1 / 2) return hubble if __name__ == "__main__": import doctest # run doctest doctest.testmod() # demo LCDM approximation _snake_case = 0.3 print( hubble_parameter( hubble_constant=68.3, radiation_density=1e-4, matter_density=matter_density, dark_energy=1 - matter_density, redshift=0, ) )
294
1
"""simple docstring""" # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from accelerate import PartialState from accelerate.utils.operations import broadcast, gather, gather_object, pad_across_processes, reduce def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' return (torch.arange(state.num_processes ) + 1.0 + (state.num_processes * state.process_index)).to(state.device ) def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : List[Any] = create_tensor(UpperCamelCase__ ) _a : Optional[Any] = gather(UpperCamelCase__ ) assert gathered_tensor.tolist() == list(range(1 , state.num_processes**2 + 1 ) ) def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : Dict = [state.process_index] _a : List[Any] = gather_object(UpperCamelCase__ ) assert len(UpperCamelCase__ ) == state.num_processes, F"""{gathered_obj}, {len(UpperCamelCase__ )} != {state.num_processes}""" assert gathered_obj == list(range(state.num_processes ) ), F"""{gathered_obj} != {list(range(state.num_processes ) )}""" def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : Any = create_tensor(UpperCamelCase__ ) _a : Optional[Any] = broadcast(UpperCamelCase__ ) assert broadcasted_tensor.shape == torch.Size([state.num_processes] ) assert broadcasted_tensor.tolist() == list(range(1 , state.num_processes + 1 ) ) def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' # We need to pad the tensor with one more element if we are the main process # to ensure that we can pad if state.is_main_process: _a : Optional[int] = torch.arange(state.num_processes + 1 ).to(state.device ) else: _a : Dict = torch.arange(state.num_processes ).to(state.device ) _a : Union[str, Any] = pad_across_processes(UpperCamelCase__ ) assert padded_tensor.shape == torch.Size([state.num_processes + 1] ) if not state.is_main_process: assert padded_tensor.tolist() == list(range(0 , state.num_processes ) ) + [0] def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' # For now runs on only two processes if state.num_processes != 2: return _a : List[str] = create_tensor(UpperCamelCase__ ) _a : int = reduce(UpperCamelCase__ , """sum""" ) _a : Tuple = torch.tensor([4.0, 6] ).to(state.device ) assert torch.allclose(UpperCamelCase__ , UpperCamelCase__ ), F"""{reduced_tensor} != {truth_tensor}""" def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' # For now runs on only two processes if state.num_processes != 2: return _a : Optional[int] = create_tensor(UpperCamelCase__ ) _a : Union[str, Any] = reduce(UpperCamelCase__ , """mean""" ) _a : List[Any] = torch.tensor([2.0, 3] ).to(state.device ) assert torch.allclose(UpperCamelCase__ , UpperCamelCase__ ), F"""{reduced_tensor} != {truth_tensor}""" def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' # For xla_spawn (TPUs) main() def lowerCAmelCase__ ( ): '''simple docstring''' _a : List[Any] = PartialState() state.print(F"""State: {state}""" ) state.print("""testing gather""" ) test_gather(UpperCamelCase__ ) state.print("""testing gather_object""" ) test_gather_object(UpperCamelCase__ ) state.print("""testing broadcast""" ) test_broadcast(UpperCamelCase__ ) state.print("""testing pad_across_processes""" ) test_pad_across_processes(UpperCamelCase__ ) state.print("""testing reduce_sum""" ) test_reduce_sum(UpperCamelCase__ ) state.print("""testing reduce_mean""" ) test_reduce_mean(UpperCamelCase__ ) if __name__ == "__main__": main()
294
"""simple docstring""" import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : List[Any] ) -> Dict: _a : Optional[int] = tempfile.mkdtemp() _a : Optional[Any] = SamImageProcessor() _a : int = SamProcessor(UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) def _lowercase ( self : Tuple , **UpperCAmelCase__ : Any ) -> Any: return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase__ ).image_processor def _lowercase ( self : str ) -> int: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Tuple ) -> Dict: _a : List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _a : Optional[int] = [Image.fromarray(np.moveaxis(UpperCAmelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def _lowercase ( self : Dict ) -> Dict: _a : List[Any] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _a : Tuple = self.get_image_processor(do_normalize=UpperCAmelCase__ , padding_value=1.0 ) _a : Tuple = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=UpperCAmelCase__ , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCAmelCase__ ) def _lowercase ( self : Union[str, Any] ) -> Tuple: _a : Optional[Any] = self.get_image_processor() _a : int = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Union[str, Any] = self.prepare_image_inputs() _a : List[str] = image_processor(UpperCAmelCase__ , return_tensors="""np""" ) _a : List[str] = processor(images=UpperCAmelCase__ , return_tensors="""np""" ) input_feat_extract.pop("""original_sizes""" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("""reshaped_input_sizes""" ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_torch def _lowercase ( self : Optional[Any] ) -> Optional[Any]: _a : Optional[Any] = self.get_image_processor() _a : Dict = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Tuple = [torch.ones((1, 3, 5, 5) )] _a : Tuple = [[1764, 2646]] _a : Optional[int] = [[683, 1024]] _a : List[Any] = processor.post_process_masks(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) _a : int = processor.post_process_masks( UpperCAmelCase__ , torch.tensor(UpperCAmelCase__ ) , torch.tensor(UpperCAmelCase__ ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np _a : Optional[Any] = [np.ones((1, 3, 5, 5) )] _a : Tuple = processor.post_process_masks(UpperCAmelCase__ , np.array(UpperCAmelCase__ ) , np.array(UpperCAmelCase__ ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) _a : List[str] = [[1, 0], [0, 1]] with self.assertRaises(UpperCAmelCase__ ): _a : str = processor.post_process_masks(UpperCAmelCase__ , np.array(UpperCAmelCase__ ) , np.array(UpperCAmelCase__ ) ) @require_vision @require_tf class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : Any ) -> List[str]: _a : List[str] = tempfile.mkdtemp() _a : Any = SamImageProcessor() _a : Union[str, Any] = SamProcessor(UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) def _lowercase ( self : List[str] , **UpperCAmelCase__ : Any ) -> List[str]: return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase__ ).image_processor def _lowercase ( self : Optional[Any] ) -> Union[str, Any]: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Dict ) -> List[str]: _a : List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _a : List[Any] = [Image.fromarray(np.moveaxis(UpperCAmelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def _lowercase ( self : Union[str, Any] ) -> Union[str, Any]: _a : Optional[int] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _a : str = self.get_image_processor(do_normalize=UpperCAmelCase__ , padding_value=1.0 ) _a : Union[str, Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=UpperCAmelCase__ , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCAmelCase__ ) def _lowercase ( self : List[Any] ) -> str: _a : Union[str, Any] = self.get_image_processor() _a : Dict = SamProcessor(image_processor=UpperCAmelCase__ ) _a : int = self.prepare_image_inputs() _a : List[str] = image_processor(UpperCAmelCase__ , return_tensors="""np""" ) _a : List[str] = processor(images=UpperCAmelCase__ , return_tensors="""np""" ) input_feat_extract.pop("""original_sizes""" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("""reshaped_input_sizes""" ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_tf def _lowercase ( self : Optional[Any] ) -> int: _a : Optional[Any] = self.get_image_processor() _a : Dict = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Any = [tf.ones((1, 3, 5, 5) )] _a : Tuple = [[1764, 2646]] _a : str = [[683, 1024]] _a : Union[str, Any] = processor.post_process_masks(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , return_tensors="""tf""" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) _a : Union[str, Any] = processor.post_process_masks( UpperCAmelCase__ , tf.convert_to_tensor(UpperCAmelCase__ ) , tf.convert_to_tensor(UpperCAmelCase__ ) , return_tensors="""tf""" , ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np _a : List[Any] = [np.ones((1, 3, 5, 5) )] _a : Optional[int] = processor.post_process_masks( UpperCAmelCase__ , np.array(UpperCAmelCase__ ) , np.array(UpperCAmelCase__ ) , return_tensors="""tf""" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) _a : Dict = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): _a : List[Any] = processor.post_process_masks( UpperCAmelCase__ , np.array(UpperCAmelCase__ ) , np.array(UpperCAmelCase__ ) , return_tensors="""tf""" ) @require_vision @require_torchvision class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : str ) -> Optional[Any]: _a : Optional[Any] = tempfile.mkdtemp() _a : Dict = SamImageProcessor() _a : List[str] = SamProcessor(UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) def _lowercase ( self : Any , **UpperCAmelCase__ : Dict ) -> int: return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase__ ).image_processor def _lowercase ( self : Tuple ) -> List[Any]: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : str ) -> int: _a : str = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _a : int = [Image.fromarray(np.moveaxis(UpperCAmelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def _lowercase ( self : int ) -> List[Any]: _a : Optional[Any] = self.get_image_processor() _a : Optional[Any] = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Tuple = np.random.randint(0 , 2 , size=(1, 3, 5, 5) ).astype(np.floataa ) _a : str = [tf.convert_to_tensor(UpperCAmelCase__ )] _a : Optional[int] = [torch.tensor(UpperCAmelCase__ )] _a : Union[str, Any] = [[1764, 2646]] _a : List[str] = [[683, 1024]] _a : Optional[int] = processor.post_process_masks( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , return_tensors="""tf""" ) _a : List[str] = processor.post_process_masks( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , return_tensors="""pt""" ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def _lowercase ( self : str ) -> Optional[Any]: _a : List[Any] = self.get_image_processor() _a : Any = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Dict = self.prepare_image_inputs() _a : List[str] = image_processor(UpperCAmelCase__ , return_tensors="""pt""" )["""pixel_values"""].numpy() _a : str = processor(images=UpperCAmelCase__ , return_tensors="""pt""" )["""pixel_values"""].numpy() _a : Optional[Any] = image_processor(UpperCAmelCase__ , return_tensors="""tf""" )["""pixel_values"""].numpy() _a : Optional[int] = processor(images=UpperCAmelCase__ , return_tensors="""tf""" )["""pixel_values"""].numpy() self.assertTrue(np.allclose(UpperCAmelCase__ , UpperCAmelCase__ ) ) self.assertTrue(np.allclose(UpperCAmelCase__ , UpperCAmelCase__ ) ) self.assertTrue(np.allclose(UpperCAmelCase__ , UpperCAmelCase__ ) )
294
1
"""simple docstring""" from __future__ import annotations _snake_case = list[list[int]] # assigning initial values to the grid _snake_case = [ [3, 0, 6, 5, 0, 8, 4, 0, 0], [5, 2, 0, 0, 0, 0, 0, 0, 0], [0, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] # a grid with no solution _snake_case = [ [5, 0, 6, 5, 0, 8, 4, 0, 3], [5, 2, 0, 0, 0, 0, 0, 0, 2], [1, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' for i in range(9 ): if grid[row][i] == n or grid[i][column] == n: return False for i in range(3 ): for j in range(3 ): if grid[(row - row % 3) + i][(column - column % 3) + j] == n: return False return True def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' for i in range(9 ): for j in range(9 ): if grid[i][j] == 0: return i, j return None def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' if location := find_empty_location(UpperCamelCase__ ): _a , _a : List[str] = location else: # If the location is ``None``, then the grid is solved. return grid for digit in range(1 , 1_0 ): if is_safe(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): _a : List[str] = digit if sudoku(UpperCamelCase__ ) is not None: return grid _a : Optional[int] = 0 return None def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' for row in grid: for cell in row: print(UpperCamelCase__ , end=""" """ ) print() if __name__ == "__main__": # make a copy of grid so that you can compare with the unmodified grid for example_grid in (initial_grid, no_solution): print('\nExample grid:\n' + '=' * 20) print_solution(example_grid) print('\nExample grid solution:') _snake_case = sudoku(example_grid) if solution is not None: print_solution(solution) else: print('Cannot find a solution.')
294
"""simple docstring""" import argparse import gc import json import os import re import torch from huggingface_hub import hf_hub_download from transformers import AutoModelForCausalLM, AutoTokenizer, PreTrainedTokenizerFast, RwkvConfig from transformers.modeling_utils import WEIGHTS_INDEX_NAME, shard_checkpoint _snake_case = { '169M': 12, '430M': 24, '1B5': 24, '3B': 32, '7B': 32, '14B': 40, } _snake_case = { '169M': 768, '430M': 1024, '1B5': 2048, '3B': 2560, '7B': 4096, '14B': 5120, } def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : int = list(state_dict.keys() ) for name in state_dict_keys: _a : str = state_dict.pop(UpperCamelCase__ ) # emb -> embedding if name.startswith("""emb.""" ): _a : Dict = name.replace("""emb.""" , """embeddings.""" ) # ln_0 -> pre_ln (only present at block 0) if name.startswith("""blocks.0.ln0""" ): _a : Dict = name.replace("""blocks.0.ln0""" , """blocks.0.pre_ln""" ) # att -> attention _a : Any = re.sub(R"""blocks\.(\d+)\.att""" , R"""blocks.\1.attention""" , UpperCamelCase__ ) # ffn -> feed_forward _a : int = re.sub(R"""blocks\.(\d+)\.ffn""" , R"""blocks.\1.feed_forward""" , UpperCamelCase__ ) # time_mix_k -> time_mix_key and reshape if name.endswith(""".time_mix_k""" ): _a : List[str] = name.replace(""".time_mix_k""" , """.time_mix_key""" ) # time_mix_v -> time_mix_value and reshape if name.endswith(""".time_mix_v""" ): _a : Tuple = name.replace(""".time_mix_v""" , """.time_mix_value""" ) # time_mix_r -> time_mix_key and reshape if name.endswith(""".time_mix_r""" ): _a : Dict = name.replace(""".time_mix_r""" , """.time_mix_receptance""" ) if name != "head.weight": _a : Optional[int] = """rwkv.""" + name _a : Any = weight return state_dict def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=False , UpperCamelCase__=None ): '''simple docstring''' # 1. If possible, build the tokenizer. if tokenizer_file is None: print("""No `--tokenizer_file` provided, we will use the default tokenizer.""" ) _a : Tuple = 5_0_2_7_7 _a : str = AutoTokenizer.from_pretrained("""EleutherAI/gpt-neox-20b""" ) else: _a : int = PreTrainedTokenizerFast(tokenizer_file=UpperCamelCase__ ) _a : int = len(UpperCamelCase__ ) tokenizer.save_pretrained(UpperCamelCase__ ) # 2. Build the config _a : Optional[Any] = list(NUM_HIDDEN_LAYERS_MAPPING.keys() ) if size is None: # Try to infer size from the checkpoint name for candidate in possible_sizes: if candidate in checkpoint_file: _a : Tuple = candidate break if size is None: raise ValueError("""Could not infer the size, please provide it with the `--size` argument.""" ) if size not in possible_sizes: raise ValueError(F"""`size` should be one of {possible_sizes}, got {size}.""" ) _a : List[Any] = RwkvConfig( vocab_size=UpperCamelCase__ , num_hidden_layers=NUM_HIDDEN_LAYERS_MAPPING[size] , hidden_size=HIDEN_SIZE_MAPPING[size] , ) config.save_pretrained(UpperCamelCase__ ) # 3. Download model file then convert state_dict _a : str = hf_hub_download(UpperCamelCase__ , UpperCamelCase__ ) _a : int = torch.load(UpperCamelCase__ , map_location="""cpu""" ) _a : List[str] = convert_state_dict(UpperCamelCase__ ) # 4. Split in shards and save _a , _a : List[str] = shard_checkpoint(UpperCamelCase__ ) for shard_file, shard in shards.items(): torch.save(UpperCamelCase__ , os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) if index is not None: _a : Optional[Any] = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) # Save the index as well with open(UpperCamelCase__ , """w""" , encoding="""utf-8""" ) as f: _a : Dict = json.dumps(UpperCamelCase__ , indent=2 , sort_keys=UpperCamelCase__ ) + """\n""" f.write(UpperCamelCase__ ) # 5. Clean up shards (for some reason the file PyTorch saves take the same space as the whole state_dict print( """Cleaning up shards. This may error with an OOM error, it this is the case don't worry you still have converted the model.""" ) _a : List[Any] = list(shards.keys() ) del state_dict del shards gc.collect() for shard_file in shard_files: _a : Any = torch.load(os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) torch.save({k: v.cpu().clone() for k, v in state_dict.items()} , os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) del state_dict gc.collect() if push_to_hub: if model_name is None: raise ValueError("""Please provide a `model_name` to push the model to the Hub.""" ) _a : Dict = AutoModelForCausalLM.from_pretrained(UpperCamelCase__ ) model.push_to_hub(UpperCamelCase__ , max_shard_size="""2GB""" ) tokenizer.push_to_hub(UpperCamelCase__ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '--repo_id', default=None, type=str, required=True, help='Repo ID from which to pull the checkpoint.' ) parser.add_argument( '--checkpoint_file', default=None, type=str, required=True, help='Name of the checkpoint file in the repo.' ) parser.add_argument( '--output_dir', default=None, type=str, required=True, help='Where to save the converted model.' ) parser.add_argument( '--tokenizer_file', default=None, type=str, help='Path to the tokenizer file to use (if not provided, only the model is converted).', ) parser.add_argument( '--size', default=None, type=str, help='Size of the model. Will be inferred from the `checkpoint_file` if not passed.', ) parser.add_argument( '--push_to_hub', action='store_true', help='Push to the Hub the converted model.', ) parser.add_argument( '--model_name', default=None, type=str, help='Name of the pushed model on the Hub, including the username / organization.', ) _snake_case = parser.parse_args() convert_rmkv_checkpoint_to_hf_format( args.repo_id, args.checkpoint_file, args.output_dir, size=args.size, tokenizer_file=args.tokenizer_file, push_to_hub=args.push_to_hub, model_name=args.model_name, )
294
1
"""simple docstring""" import json import os import tempfile from transformers.testing_utils import check_json_file_has_correct_format class UpperCamelCase : UpperCamelCase : Union[str, Any] = None def _lowercase ( self : Optional[int] ) -> Any: _a : Optional[int] = self.feature_extraction_class(**self.feat_extract_dict ) _a : Any = json.loads(feat_extract.to_json_string() ) for key, value in self.feat_extract_dict.items(): self.assertEqual(obj[key] , UpperCAmelCase__ ) def _lowercase ( self : Dict ) -> Dict: _a : Optional[int] = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _a : Any = os.path.join(UpperCAmelCase__ , """feat_extract.json""" ) feat_extract_first.to_json_file(UpperCAmelCase__ ) _a : Optional[int] = self.feature_extraction_class.from_json_file(UpperCAmelCase__ ) self.assertEqual(feat_extract_second.to_dict() , feat_extract_first.to_dict() ) def _lowercase ( self : Union[str, Any] ) -> str: _a : Optional[Any] = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _a : Optional[Any] = feat_extract_first.save_pretrained(UpperCAmelCase__ )[0] check_json_file_has_correct_format(UpperCAmelCase__ ) _a : Optional[int] = self.feature_extraction_class.from_pretrained(UpperCAmelCase__ ) self.assertEqual(feat_extract_second.to_dict() , feat_extract_first.to_dict() ) def _lowercase ( self : List[str] ) -> Optional[Any]: _a : Union[str, Any] = self.feature_extraction_class() self.assertIsNotNone(UpperCAmelCase__ )
294
"""simple docstring""" import shutil import tempfile import unittest from transformers import ClapFeatureExtractor, ClapProcessor, RobertaTokenizer, RobertaTokenizerFast from transformers.testing_utils import require_sentencepiece, require_torchaudio from .test_feature_extraction_clap import floats_list @require_torchaudio @require_sentencepiece class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : int ) -> List[str]: _a : Any = """laion/clap-htsat-unfused""" _a : Union[str, Any] = tempfile.mkdtemp() def _lowercase ( self : List[Any] , **UpperCAmelCase__ : Any ) -> Dict: return RobertaTokenizer.from_pretrained(self.checkpoint , **UpperCAmelCase__ ) def _lowercase ( self : List[Any] , **UpperCAmelCase__ : List[str] ) -> int: return ClapFeatureExtractor.from_pretrained(self.checkpoint , **UpperCAmelCase__ ) def _lowercase ( self : List[Any] ) -> Tuple: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : List[str] ) -> Optional[int]: _a : List[str] = self.get_tokenizer() _a : Any = self.get_feature_extractor() _a : Optional[Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) _a : List[str] = ClapProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCAmelCase__ ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , UpperCAmelCase__ ) def _lowercase ( self : Dict ) -> Optional[int]: _a : Tuple = ClapProcessor(tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() ) processor.save_pretrained(self.tmpdirname ) _a : Dict = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) _a : Union[str, Any] = self.get_feature_extractor(do_normalize=UpperCAmelCase__ , padding_value=1.0 ) _a : Union[str, Any] = ClapProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=UpperCAmelCase__ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCAmelCase__ ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.feature_extractor , UpperCAmelCase__ ) def _lowercase ( self : List[str] ) -> Optional[Any]: _a : Optional[int] = self.get_feature_extractor() _a : Tuple = self.get_tokenizer() _a : List[Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) _a : Any = floats_list((3, 1000) ) _a : List[Any] = feature_extractor(UpperCAmelCase__ , return_tensors="""np""" ) _a : List[str] = processor(audios=UpperCAmelCase__ , return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def _lowercase ( self : Tuple ) -> Optional[int]: _a : List[str] = self.get_feature_extractor() _a : Any = self.get_tokenizer() _a : Any = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) _a : Optional[int] = """This is a test string""" _a : Tuple = processor(text=UpperCAmelCase__ ) _a : int = tokenizer(UpperCAmelCase__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _lowercase ( self : List[Any] ) -> Any: _a : str = self.get_feature_extractor() _a : List[str] = self.get_tokenizer() _a : List[Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) _a : Any = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _a : Dict = processor.batch_decode(UpperCAmelCase__ ) _a : Any = tokenizer.batch_decode(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def _lowercase ( self : Dict ) -> List[str]: _a : str = self.get_feature_extractor() _a : Optional[Any] = self.get_tokenizer() _a : Union[str, Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) self.assertListEqual( processor.model_input_names[2:] , feature_extractor.model_input_names , msg="""`processor` and `feature_extractor` model input names do not match""" , )
294
1
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { 'asapp/sew-d-tiny-100k': 'https://huggingface.co/asapp/sew-d-tiny-100k/resolve/main/config.json', # See all SEW-D models at https://huggingface.co/models?filter=sew-d } class UpperCamelCase ( snake_case_ ): UpperCamelCase : Dict = '''sew-d''' def __init__( self : str , UpperCAmelCase__ : str=32 , UpperCAmelCase__ : Tuple=768 , UpperCAmelCase__ : int=12 , UpperCAmelCase__ : List[Any]=12 , UpperCAmelCase__ : int=3072 , UpperCAmelCase__ : Dict=2 , UpperCAmelCase__ : int=512 , UpperCAmelCase__ : int=256 , UpperCAmelCase__ : Tuple=True , UpperCAmelCase__ : Any=True , UpperCAmelCase__ : List[str]=("p2c", "c2p") , UpperCAmelCase__ : Tuple="layer_norm" , UpperCAmelCase__ : Union[str, Any]="gelu_python" , UpperCAmelCase__ : List[Any]=0.1 , UpperCAmelCase__ : Any=0.1 , UpperCAmelCase__ : Optional[int]=0.1 , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : List[Any]=0.1 , UpperCAmelCase__ : Optional[Any]=0.0_2 , UpperCAmelCase__ : Any=1E-7 , UpperCAmelCase__ : Dict=1E-5 , UpperCAmelCase__ : Dict="group" , UpperCAmelCase__ : Optional[int]="gelu" , UpperCAmelCase__ : Dict=(64, 128, 128, 128, 128, 256, 256, 256, 256, 512, 512, 512, 512) , UpperCAmelCase__ : Optional[Any]=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , UpperCAmelCase__ : Tuple=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , UpperCAmelCase__ : Dict=False , UpperCAmelCase__ : Optional[int]=128 , UpperCAmelCase__ : str=16 , UpperCAmelCase__ : Optional[Any]=True , UpperCAmelCase__ : Any=0.0_5 , UpperCAmelCase__ : List[str]=10 , UpperCAmelCase__ : List[Any]=2 , UpperCAmelCase__ : List[Any]=0.0 , UpperCAmelCase__ : Dict=10 , UpperCAmelCase__ : Dict=0 , UpperCAmelCase__ : Any="mean" , UpperCAmelCase__ : Optional[int]=False , UpperCAmelCase__ : str=False , UpperCAmelCase__ : str=256 , UpperCAmelCase__ : Union[str, Any]=0 , UpperCAmelCase__ : Optional[int]=1 , UpperCAmelCase__ : Optional[Any]=2 , **UpperCAmelCase__ : Tuple , ) -> Optional[int]: super().__init__(**UpperCAmelCase__ , pad_token_id=UpperCAmelCase__ , bos_token_id=UpperCAmelCase__ , eos_token_id=UpperCAmelCase__ ) _a : int = hidden_size _a : Optional[int] = feat_extract_norm _a : Union[str, Any] = feat_extract_activation _a : Union[str, Any] = list(UpperCAmelCase__ ) _a : str = list(UpperCAmelCase__ ) _a : Tuple = list(UpperCAmelCase__ ) _a : Union[str, Any] = conv_bias _a : Optional[Any] = num_conv_pos_embeddings _a : Any = num_conv_pos_embedding_groups _a : Any = len(self.conv_dim ) _a : str = num_hidden_layers _a : Dict = intermediate_size _a : Optional[Any] = squeeze_factor _a : Tuple = max_position_embeddings _a : int = position_buckets _a : Optional[int] = share_att_key _a : Any = relative_attention _a : str = norm_rel_ebd _a : Dict = list(UpperCAmelCase__ ) _a : List[Any] = hidden_act _a : Tuple = num_attention_heads _a : Optional[int] = hidden_dropout _a : Optional[Any] = attention_dropout _a : str = activation_dropout _a : Optional[Any] = feat_proj_dropout _a : str = final_dropout _a : Optional[int] = layer_norm_eps _a : List[str] = feature_layer_norm_eps _a : Dict = initializer_range _a : int = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect.""" """It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,""" f"""but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)""" f"""= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`.""" ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 _a : List[str] = apply_spec_augment _a : Any = mask_time_prob _a : int = mask_time_length _a : int = mask_time_min_masks _a : Dict = mask_feature_prob _a : int = mask_feature_length _a : str = mask_feature_min_masks # ctc loss _a : Union[str, Any] = ctc_loss_reduction _a : int = ctc_zero_infinity # sequence classification _a : int = use_weighted_layer_sum _a : str = classifier_proj_size @property def _lowercase ( self : List[str] ) -> int: return functools.reduce(operator.mul , self.conv_stride , 1 )
294
"""simple docstring""" import warnings from ...utils import logging from .image_processing_clip import CLIPImageProcessor _snake_case = logging.get_logger(__name__) class UpperCamelCase ( snake_case_ ): def __init__( self : Any , *UpperCAmelCase__ : Optional[Any] , **UpperCAmelCase__ : Tuple ) -> None: warnings.warn( """The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use CLIPImageProcessor instead.""" , UpperCAmelCase__ , ) super().__init__(*UpperCAmelCase__ , **UpperCAmelCase__ )
294
1
"""simple docstring""" import torch from torch import nn class UpperCamelCase ( nn.Module ): def __init__( self : List[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Any=1 , UpperCAmelCase__ : Any=False ) -> Union[str, Any]: super().__init__() _a : Union[str, Any] = n_token _a : Optional[Any] = d_embed _a : Optional[Any] = d_proj _a : Dict = cutoffs + [n_token] _a : Union[str, Any] = [0] + self.cutoffs _a : Any = div_val _a : List[str] = self.cutoffs[0] _a : List[str] = len(self.cutoffs ) - 1 _a : Optional[Any] = self.shortlist_size + self.n_clusters if self.n_clusters > 0: _a : int = nn.Parameter(torch.zeros(self.n_clusters , self.d_embed ) ) _a : Dict = nn.Parameter(torch.zeros(self.n_clusters ) ) _a : Optional[Any] = nn.ModuleList() _a : str = nn.ParameterList() if div_val == 1: for i in range(len(self.cutoffs ) ): if d_proj != d_embed: self.out_projs.append(nn.Parameter(torch.FloatTensor(UpperCAmelCase__ , UpperCAmelCase__ ) ) ) else: self.out_projs.append(UpperCAmelCase__ ) self.out_layers.append(nn.Linear(UpperCAmelCase__ , UpperCAmelCase__ ) ) else: for i in range(len(self.cutoffs ) ): _a , _a : Union[str, Any] = self.cutoff_ends[i], self.cutoff_ends[i + 1] _a : Optional[Any] = d_embed // (div_val**i) self.out_projs.append(nn.Parameter(torch.FloatTensor(UpperCAmelCase__ , UpperCAmelCase__ ) ) ) self.out_layers.append(nn.Linear(UpperCAmelCase__ , r_idx - l_idx ) ) _a : List[Any] = keep_order def _lowercase ( self : Tuple , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : List[Any] ) -> List[str]: if proj is None: _a : Optional[Any] = nn.functional.linear(UpperCAmelCase__ , UpperCAmelCase__ , bias=UpperCAmelCase__ ) else: # if CUDA_MAJOR <= 9 and CUDA_MINOR <= 1: _a : str = nn.functional.linear(UpperCAmelCase__ , proj.t().contiguous() ) _a : Tuple = nn.functional.linear(UpperCAmelCase__ , UpperCAmelCase__ , bias=UpperCAmelCase__ ) # else: # logit = torch.einsum('bd,de,ev->bv', (hidden, proj, weight.t())) # if bias is not None: # logit = logit + bias return logit def _lowercase ( self : Optional[int] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Optional[Any]=False ) -> Dict: if labels is not None: # Shift so that tokens < n predict n _a : Dict = hidden[..., :-1, :].contiguous() _a : Dict = labels[..., 1:].contiguous() _a : Optional[Any] = hidden.view(-1 , hidden.size(-1 ) ) _a : Union[str, Any] = labels.view(-1 ) if hidden.size(0 ) != labels.size(0 ): raise RuntimeError("""Input and labels should have the same size in the batch dimension.""" ) else: _a : Optional[int] = hidden.view(-1 , hidden.size(-1 ) ) if self.n_clusters == 0: _a : Any = self._compute_logit(UpperCAmelCase__ , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0] ) if labels is not None: _a : Dict = labels != -100 _a : Tuple = torch.zeros_like(UpperCAmelCase__ , dtype=hidden.dtype , device=hidden.device ) _a : List[Any] = ( -nn.functional.log_softmax(UpperCAmelCase__ , dim=-1 )[mask].gather(1 , labels[mask].unsqueeze(1 ) ).squeeze(1 ) ) else: _a : Tuple = nn.functional.log_softmax(UpperCAmelCase__ , dim=-1 ) else: # construct weights and biases _a , _a : Dict = [], [] for i in range(len(self.cutoffs ) ): if self.div_val == 1: _a , _a : Optional[Any] = self.cutoff_ends[i], self.cutoff_ends[i + 1] _a : Tuple = self.out_layers[0].weight[l_idx:r_idx] _a : Any = self.out_layers[0].bias[l_idx:r_idx] else: _a : Tuple = self.out_layers[i].weight _a : Dict = self.out_layers[i].bias if i == 0: _a : Dict = torch.cat([weight_i, self.cluster_weight] , dim=0 ) _a : List[Any] = torch.cat([bias_i, self.cluster_bias] , dim=0 ) weights.append(UpperCAmelCase__ ) biases.append(UpperCAmelCase__ ) _a , _a , _a : List[Any] = weights[0], biases[0], self.out_projs[0] _a : Optional[int] = self._compute_logit(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) _a : str = nn.functional.log_softmax(UpperCAmelCase__ , dim=1 ) if labels is None: _a : int = hidden.new_empty((head_logit.size(0 ), self.n_token) ) else: _a : List[Any] = torch.zeros_like(UpperCAmelCase__ , dtype=hidden.dtype , device=hidden.device ) _a : List[Any] = 0 _a : Any = [0] + self.cutoffs for i in range(len(UpperCAmelCase__ ) - 1 ): _a , _a : Tuple = cutoff_values[i], cutoff_values[i + 1] if labels is not None: _a : List[str] = (labels >= l_idx) & (labels < r_idx) _a : List[str] = mask_i.nonzero().squeeze() if indices_i.numel() == 0: continue _a : Optional[Any] = labels.index_select(0 , UpperCAmelCase__ ) - l_idx _a : str = head_logprob.index_select(0 , UpperCAmelCase__ ) _a : Dict = hidden.index_select(0 , UpperCAmelCase__ ) else: _a : Tuple = hidden if i == 0: if labels is not None: _a : Optional[int] = head_logprob_i.gather(1 , target_i[:, None] ).squeeze(1 ) else: _a : Optional[Any] = head_logprob[:, : self.cutoffs[0]] else: _a , _a , _a : int = weights[i], biases[i], self.out_projs[i] _a : Optional[int] = self._compute_logit(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) _a : Any = nn.functional.log_softmax(UpperCAmelCase__ , dim=1 ) _a : Optional[int] = self.cutoffs[0] + i - 1 # No probability for the head cluster if labels is not None: _a : Optional[int] = head_logprob_i[:, cluster_prob_idx] + tail_logprob_i.gather( 1 , target_i[:, None] ).squeeze(1 ) else: _a : str = head_logprob[:, cluster_prob_idx, None] + tail_logprob_i _a : Tuple = logprob_i if labels is not None: if (hasattr(self , """keep_order""" ) and self.keep_order) or keep_order: out.index_copy_(0 , UpperCAmelCase__ , -logprob_i ) else: out[offset : offset + logprob_i.size(0 )].copy_(-logprob_i ) offset += logprob_i.size(0 ) return out def _lowercase ( self : int , UpperCAmelCase__ : Tuple ) -> int: if self.n_clusters == 0: _a : Tuple = self._compute_logit(UpperCAmelCase__ , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0] ) return nn.functional.log_softmax(UpperCAmelCase__ , dim=-1 ) else: # construct weights and biases _a , _a : List[str] = [], [] for i in range(len(self.cutoffs ) ): if self.div_val == 1: _a , _a : List[str] = self.cutoff_ends[i], self.cutoff_ends[i + 1] _a : Optional[int] = self.out_layers[0].weight[l_idx:r_idx] _a : Any = self.out_layers[0].bias[l_idx:r_idx] else: _a : Optional[int] = self.out_layers[i].weight _a : Optional[int] = self.out_layers[i].bias if i == 0: _a : List[str] = torch.cat([weight_i, self.cluster_weight] , dim=0 ) _a : Dict = torch.cat([bias_i, self.cluster_bias] , dim=0 ) weights.append(UpperCAmelCase__ ) biases.append(UpperCAmelCase__ ) _a , _a , _a : Tuple = weights[0], biases[0], self.out_projs[0] _a : Dict = self._compute_logit(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) _a : Union[str, Any] = hidden.new_empty((head_logit.size(0 ), self.n_token) ) _a : Optional[Any] = nn.functional.log_softmax(UpperCAmelCase__ , dim=1 ) _a : Optional[int] = [0] + self.cutoffs for i in range(len(UpperCAmelCase__ ) - 1 ): _a , _a : Tuple = cutoff_values[i], cutoff_values[i + 1] if i == 0: _a : Any = head_logprob[:, : self.cutoffs[0]] else: _a , _a , _a : Optional[Any] = weights[i], biases[i], self.out_projs[i] _a : Union[str, Any] = self._compute_logit(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) _a : Dict = nn.functional.log_softmax(UpperCAmelCase__ , dim=1 ) _a : List[Any] = head_logprob[:, -i] + tail_logprob_i _a : Optional[Any] = logprob_i return out
294
"""simple docstring""" import unittest import numpy as np def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = None , ): '''simple docstring''' _a : List[Any] = np.shape(UpperCamelCase__ ) _a : Any = np.shape(UpperCamelCase__ ) _a : Union[str, Any] = np.shape(UpperCamelCase__ ) if shape_a[0] != shape_b[0]: _a : int = ( """Expected the same number of rows for A and B. """ F"""Instead found A of size {shape_a} and B of size {shape_b}""" ) raise ValueError(UpperCamelCase__ ) if shape_b[1] != shape_c[1]: _a : Tuple = ( """Expected the same number of columns for B and C. """ F"""Instead found B of size {shape_b} and C of size {shape_c}""" ) raise ValueError(UpperCamelCase__ ) _a : int = pseudo_inv if a_inv is None: try: _a : Optional[int] = np.linalg.inv(UpperCamelCase__ ) except np.linalg.LinAlgError: raise ValueError( """Input matrix A is not invertible. Cannot compute Schur complement.""" ) return mat_c - mat_b.T @ a_inv @ mat_b class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : int ) -> None: _a : str = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) _a : Tuple = np.array([[0, 3], [3, 0], [2, 3]] ) _a : Optional[int] = np.array([[2, 1], [6, 3]] ) _a : Optional[Any] = schur_complement(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) _a : Union[str, Any] = np.block([[a, b], [b.T, c]] ) _a : int = np.linalg.det(UpperCAmelCase__ ) _a : Union[str, Any] = np.linalg.det(UpperCAmelCase__ ) _a : List[Any] = np.linalg.det(UpperCAmelCase__ ) self.assertAlmostEqual(UpperCAmelCase__ , det_a * det_s ) def _lowercase ( self : int ) -> None: _a : Optional[int] = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) _a : Optional[int] = np.array([[0, 3], [3, 0], [2, 3]] ) _a : Union[str, Any] = np.array([[2, 1], [6, 3]] ) with self.assertRaises(UpperCAmelCase__ ): schur_complement(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) def _lowercase ( self : List[Any] ) -> None: _a : Any = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) _a : Dict = np.array([[0, 3], [3, 0], [2, 3]] ) _a : List[Any] = np.array([[2, 1, 3], [6, 3, 5]] ) with self.assertRaises(UpperCAmelCase__ ): schur_complement(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) if __name__ == "__main__": import doctest doctest.testmod() unittest.main()
294
1
"""simple docstring""" from math import ceil, sqrt def lowerCAmelCase__ ( UpperCamelCase__ = 1_0_0_0_0_0_0 ): '''simple docstring''' _a : Any = 0 for outer_width in range(3 , (limit // 4) + 2 ): if outer_width**2 > limit: _a : Optional[Any] = max(ceil(sqrt(outer_width**2 - limit ) ) , 1 ) else: _a : Optional[Any] = 1 if (outer_width - hole_width_lower_bound) % 2: hole_width_lower_bound += 1 answer += (outer_width - hole_width_lower_bound - 2) // 2 + 1 return answer if __name__ == "__main__": print(F'''{solution() = }''')
294
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _snake_case = logging.get_logger(__name__) _snake_case = torch.device('cpu') def lowerCAmelCase__ ( ): '''simple docstring''' _a : int = """http://images.cocodataset.org/val2017/000000039769.jpg""" _a : Dict = Image.open(requests.get(UpperCamelCase__ , stream=UpperCamelCase__ ).raw ) return im def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' if swiftformer_name == "swiftformer_xs": return torch.tensor([-2.17_03e00, 2.11_07e00, -2.08_11e00, 8.86_85e-01, 2.43_60e-01] ) elif swiftformer_name == "swiftformer_s": return torch.tensor([3.96_36e-01, 2.34_78e-01, -1.69_63e00, -1.73_81e00, -8.63_37e-01] ) elif swiftformer_name == "swiftformer_l1": return torch.tensor([-4.27_68e-01, -4.74_29e-01, -1.08_97e00, -1.02_48e00, 3.55_23e-02] ) elif swiftformer_name == "swiftformer_l3": return torch.tensor([-2.53_30e-01, 2.42_11e-01, -6.01_85e-01, -8.27_89e-01, -6.04_46e-02] ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Any = dct.pop(UpperCamelCase__ ) _a : Dict = val def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : Tuple = [] for k in state_dict.keys(): _a : Any = k if ".pwconv" in k: _a : int = k_new.replace(""".pwconv""" , """.point_wise_conv""" ) if ".dwconv" in k: _a : List[str] = k_new.replace(""".dwconv""" , """.depth_wise_conv""" ) if ".Proj." in k: _a : Optional[int] = k_new.replace(""".Proj.""" , """.proj.""" ) if "patch_embed" in k_new: _a : Tuple = k_new.replace("""patch_embed""" , """swiftformer.patch_embed.patch_embedding""" ) if "network" in k_new: _a : int = k_new.split(""".""" ) if ls[2].isdigit(): _a : Union[str, Any] = """swiftformer.encoder.network.""" + ls[1] + """.blocks.""" + ls[2] + """.""" + """.""".join(ls[3:] ) else: _a : Tuple = k_new.replace("""network""" , """swiftformer.encoder.network""" ) rename_keys.append((k, k_new) ) return rename_keys @torch.no_grad() def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Tuple = SwiftFormerConfig() # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size _a : Optional[int] = 1_0_0_0 _a : Optional[Any] = """huggingface/label-files""" _a : Optional[Any] = """imagenet-1k-id2label.json""" _a : List[str] = json.load(open(hf_hub_download(UpperCamelCase__ , UpperCamelCase__ , repo_type="""dataset""" ) , """r""" ) ) _a : Optional[Any] = {int(UpperCamelCase__ ): v for k, v in idalabel.items()} _a : Dict = idalabel _a : Optional[int] = {v: k for k, v in idalabel.items()} # size of the architecture if swiftformer_name == "swiftformer_xs": _a : Any = [3, 3, 6, 4] _a : int = [4_8, 5_6, 1_1_2, 2_2_0] elif swiftformer_name == "swiftformer_s": _a : Any = [3, 3, 9, 6] _a : List[str] = [4_8, 6_4, 1_6_8, 2_2_4] elif swiftformer_name == "swiftformer_l1": _a : List[Any] = [4, 3, 1_0, 5] _a : Optional[int] = [4_8, 9_6, 1_9_2, 3_8_4] elif swiftformer_name == "swiftformer_l3": _a : List[Any] = [4, 4, 1_2, 6] _a : Optional[Any] = [6_4, 1_2_8, 3_2_0, 5_1_2] # load state_dict of original model, remove and rename some keys if original_ckpt: if original_ckpt.startswith("""https""" ): _a : Tuple = torch.hub.load_state_dict_from_url(UpperCamelCase__ , map_location="""cpu""" , check_hash=UpperCamelCase__ ) else: _a : Dict = torch.load(UpperCamelCase__ , map_location="""cpu""" ) _a : int = checkpoint _a : Optional[Any] = create_rename_keys(UpperCamelCase__ ) for rename_key_src, rename_key_dest in rename_keys: rename_key(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # load HuggingFace model _a : Any = SwiftFormerForImageClassification(UpperCamelCase__ ).eval() hf_model.load_state_dict(UpperCamelCase__ ) # prepare test inputs _a : Any = prepare_img() _a : Union[str, Any] = ViTImageProcessor.from_pretrained("""preprocessor_config""" ) _a : Optional[int] = processor(images=UpperCamelCase__ , return_tensors="""pt""" ) # compare outputs from both models _a : Dict = get_expected_output(UpperCamelCase__ ) _a : int = hf_model(inputs["""pixel_values"""] ).logits assert hf_logits.shape == torch.Size([1, 1_0_0_0] ) assert torch.allclose(hf_logits[0, 0:5] , UpperCamelCase__ , atol=1e-3 ) Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) print(F"""Saving model {swiftformer_name} to {pytorch_dump_folder_path}""" ) hf_model.save_pretrained(UpperCamelCase__ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '--swiftformer_name', default='swiftformer_xs', choices=['swiftformer_xs', 'swiftformer_s', 'swiftformer_l1', 'swiftformer_l3'], type=str, help='Name of the SwiftFormer model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default='./converted_outputs/', type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument('--original_ckpt', default=None, type=str, help='Path to the original model checkpoint.') _snake_case = parser.parse_args() convert_swiftformer_checkpoint(args.swiftformer_name, args.pytorch_dump_folder_path, args.original_ckpt)
294
1
"""simple docstring""" import re def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : Optional[Any] = re.compile(R"""^(\+91[\-\s]?)?[0]?(91)?[789]\d{9}$""" ) if match := re.search(UpperCamelCase__ , UpperCamelCase__ ): return match.string == phone return False if __name__ == "__main__": print(indian_phone_validator('+918827897895'))
294
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case = { 'configuration_perceiver': ['PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PerceiverConfig', 'PerceiverOnnxConfig'], 'tokenization_perceiver': ['PerceiverTokenizer'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ['PerceiverFeatureExtractor'] _snake_case = ['PerceiverImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST', 'PerceiverForImageClassificationConvProcessing', 'PerceiverForImageClassificationFourier', 'PerceiverForImageClassificationLearned', 'PerceiverForMaskedLM', 'PerceiverForMultimodalAutoencoding', 'PerceiverForOpticalFlow', 'PerceiverForSequenceClassification', 'PerceiverLayer', 'PerceiverModel', 'PerceiverPreTrainedModel', ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
294
1
"""simple docstring""" import doctest import glob import importlib import inspect import os import re from contextlib import contextmanager from functools import wraps from unittest.mock import patch import numpy as np import pytest from absl.testing import parameterized import datasets from datasets import load_metric from .utils import for_all_test_methods, local, slow # mark all tests as integration _snake_case = pytest.mark.integration _snake_case = {'comet'} _snake_case = importlib.util.find_spec('fairseq') is not None _snake_case = {'code_eval'} _snake_case = os.name == 'nt' _snake_case = {'bertscore', 'frugalscore', 'perplexity'} _snake_case = importlib.util.find_spec('transformers') is not None def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' @wraps(UpperCamelCase__ ) def wrapper(self , UpperCamelCase__ ): if not _has_fairseq and metric_name in REQUIRE_FAIRSEQ: self.skipTest("""\"test requires Fairseq\"""" ) else: test_case(self , UpperCamelCase__ ) return wrapper def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' @wraps(UpperCamelCase__ ) def wrapper(self , UpperCamelCase__ ): if not _has_transformers and metric_name in REQUIRE_TRANSFORMERS: self.skipTest("""\"test requires transformers\"""" ) else: test_case(self , UpperCamelCase__ ) return wrapper def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' @wraps(UpperCamelCase__ ) def wrapper(self , UpperCamelCase__ ): if _on_windows and metric_name in UNSUPPORTED_ON_WINDOWS: self.skipTest("""\"test not supported on Windows\"""" ) else: test_case(self , UpperCamelCase__ ) return wrapper def lowerCAmelCase__ ( ): '''simple docstring''' _a : Any = [metric_dir.split(os.sep )[-2] for metric_dir in glob.glob("""./metrics/*/""" )] return [{"testcase_name": x, "metric_name": x} for x in metrics if x != "gleu"] # gleu is unfinished @parameterized.named_parameters(get_local_metric_names() ) @for_all_test_methods( snake_case_ , snake_case_ , snake_case_ ) @local class UpperCamelCase ( parameterized.TestCase ): UpperCamelCase : Tuple = {} UpperCamelCase : List[str] = None @pytest.mark.filterwarnings("""ignore:metric_module_factory is deprecated:FutureWarning""" ) @pytest.mark.filterwarnings("""ignore:load_metric is deprecated:FutureWarning""" ) def _lowercase ( self : Union[str, Any] , UpperCAmelCase__ : Any ) -> Dict: _a : Optional[Any] = """[...]""" _a : Optional[Any] = importlib.import_module( datasets.load.metric_module_factory(os.path.join("""metrics""" , UpperCAmelCase__ ) ).module_path ) _a : List[str] = datasets.load.import_main_class(metric_module.__name__ , dataset=UpperCAmelCase__ ) # check parameters _a : Optional[Any] = inspect.signature(metric._compute ).parameters self.assertTrue(all(p.kind != p.VAR_KEYWORD for p in parameters.values() ) ) # no **kwargs # run doctest with self.patch_intensive_calls(UpperCAmelCase__ , metric_module.__name__ ): with self.use_local_metrics(): try: _a : Any = doctest.testmod(UpperCAmelCase__ , verbose=UpperCAmelCase__ , raise_on_error=UpperCAmelCase__ ) except doctest.UnexpectedException as e: raise e.exc_info[1] # raise the exception that doctest caught self.assertEqual(results.failed , 0 ) self.assertGreater(results.attempted , 1 ) @slow def _lowercase ( self : List[str] , UpperCAmelCase__ : Optional[int] ) -> int: _a : Optional[int] = """[...]""" _a : List[str] = importlib.import_module( datasets.load.metric_module_factory(os.path.join("""metrics""" , UpperCAmelCase__ ) ).module_path ) # run doctest with self.use_local_metrics(): _a : List[str] = doctest.testmod(UpperCAmelCase__ , verbose=UpperCAmelCase__ , raise_on_error=UpperCAmelCase__ ) self.assertEqual(results.failed , 0 ) self.assertGreater(results.attempted , 1 ) @contextmanager def _lowercase ( self : List[Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : int ) -> Union[str, Any]: if metric_name in self.INTENSIVE_CALLS_PATCHER: with self.INTENSIVE_CALLS_PATCHER[metric_name](UpperCAmelCase__ ): yield else: yield @contextmanager def _lowercase ( self : Tuple ) -> Optional[int]: def load_local_metric(UpperCAmelCase__ : Optional[Any] , *UpperCAmelCase__ : Union[str, Any] , **UpperCAmelCase__ : Tuple ): return load_metric(os.path.join("""metrics""" , UpperCAmelCase__ ) , *UpperCAmelCase__ , **UpperCAmelCase__ ) with patch("""datasets.load_metric""" ) as mock_load_metric: _a : Optional[int] = load_local_metric yield @classmethod def _lowercase ( cls : Union[str, Any] , UpperCAmelCase__ : Tuple ) -> List[str]: def wrapper(UpperCAmelCase__ : str ): _a : int = contextmanager(UpperCAmelCase__ ) _a : Tuple = patcher return patcher return wrapper @LocalMetricTest.register_intensive_calls_patcher("""bleurt""" ) def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' import tensorflow.compat.va as tf from bleurt.score import Predictor tf.flags.DEFINE_string("""sv""" , """""" , """""" ) # handle pytest cli flags class UpperCamelCase ( snake_case_ ): def _lowercase ( self : str , UpperCAmelCase__ : List[str] ) -> int: assert len(input_dict["""input_ids"""] ) == 2 return np.array([1.0_3, 1.0_4] ) # mock predict_fn which is supposed to do a forward pass with a bleurt model with patch("""bleurt.score._create_predictor""" ) as mock_create_predictor: _a : int = MockedPredictor() yield @LocalMetricTest.register_intensive_calls_patcher("""bertscore""" ) def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' import torch def bert_cos_score_idf(UpperCamelCase__ , UpperCamelCase__ , *UpperCamelCase__ , **UpperCamelCase__ ): return torch.tensor([[1.0, 1.0, 1.0]] * len(UpperCamelCase__ ) ) # mock get_model which is supposed to do download a bert model # mock bert_cos_score_idf which is supposed to do a forward pass with a bert model with patch("""bert_score.scorer.get_model""" ), patch( """bert_score.scorer.bert_cos_score_idf""" ) as mock_bert_cos_score_idf: _a : List[str] = bert_cos_score_idf yield @LocalMetricTest.register_intensive_calls_patcher("""comet""" ) def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' def load_from_checkpoint(UpperCamelCase__ ): class UpperCamelCase : def _lowercase ( self : Optional[int] , UpperCAmelCase__ : List[str] , *UpperCAmelCase__ : int , **UpperCAmelCase__ : Optional[int] ) -> int: assert len(UpperCAmelCase__ ) == 2 _a : str = [0.1_9, 0.9_2] return scores, sum(UpperCAmelCase__ ) / len(UpperCAmelCase__ ) return Model() # mock load_from_checkpoint which is supposed to do download a bert model # mock load_from_checkpoint which is supposed to do download a bert model with patch("""comet.download_model""" ) as mock_download_model: _a : str = None with patch("""comet.load_from_checkpoint""" ) as mock_load_from_checkpoint: _a : str = load_from_checkpoint yield def lowerCAmelCase__ ( ): '''simple docstring''' _a : Optional[Any] = load_metric(os.path.join("""metrics""" , """seqeval""" ) ) _a : Union[str, Any] = """ERROR""" _a : Union[str, Any] = F"""Scheme should be one of [IOB1, IOB2, IOE1, IOE2, IOBES, BILOU], got {wrong_scheme}""" with pytest.raises(UpperCamelCase__ , match=re.escape(UpperCamelCase__ ) ): metric.compute(predictions=[] , references=[] , scheme=UpperCamelCase__ )
294
"""simple docstring""" import cva import numpy as np class UpperCamelCase : def __init__( self : Optional[int] , UpperCAmelCase__ : float , UpperCAmelCase__ : int ) -> Dict: if k in (0.0_4, 0.0_6): _a : List[str] = k _a : List[Any] = window_size else: raise ValueError("""invalid k value""" ) def __str__( self : Dict ) -> str: return str(self.k ) def _lowercase ( self : int , UpperCAmelCase__ : str ) -> tuple[cva.Mat, list[list[int]]]: _a : Dict = cva.imread(UpperCAmelCase__ , 0 ) _a , _a : List[Any] = img.shape _a : list[list[int]] = [] _a : List[Any] = img.copy() _a : int = cva.cvtColor(UpperCAmelCase__ , cva.COLOR_GRAY2RGB ) _a , _a : Any = np.gradient(UpperCAmelCase__ ) _a : Tuple = dx**2 _a : Union[str, Any] = dy**2 _a : Union[str, Any] = dx * dy _a : int = 0.0_4 _a : List[str] = self.window_size // 2 for y in range(UpperCAmelCase__ , h - offset ): for x in range(UpperCAmelCase__ , w - offset ): _a : str = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _a : List[Any] = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _a : Tuple = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _a : Any = (wxx * wyy) - (wxy**2) _a : Tuple = wxx + wyy _a : Any = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 255 ) return color_img, corner_list if __name__ == "__main__": _snake_case = HarrisCorner(0.04, 3) _snake_case , _snake_case = edge_detect.detect('path_to_image') cva.imwrite('detect.png', color_img)
294
1
"""simple docstring""" import shutil import tempfile import unittest from transformers import ( SPIECE_UNDERLINE, AddedToken, BatchEncoding, NllbTokenizer, NllbTokenizerFast, is_torch_available, ) from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin _snake_case = get_tests_dir('fixtures/test_sentencepiece.model') if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right _snake_case = 25_6047 _snake_case = 25_6145 @require_sentencepiece @require_tokenizers class UpperCamelCase ( snake_case_ , unittest.TestCase ): UpperCamelCase : List[str] = NllbTokenizer UpperCamelCase : List[str] = NllbTokenizerFast UpperCamelCase : Optional[Any] = True UpperCamelCase : Union[str, Any] = True UpperCamelCase : Any = {} def _lowercase ( self : Any ) -> Union[str, Any]: super().setUp() # We have a SentencePiece fixture for testing _a : Tuple = NllbTokenizer(UpperCAmelCase__ , keep_accents=UpperCAmelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def _lowercase ( self : List[str] ) -> Optional[int]: _a : List[Any] = NllbTokenizer(UpperCAmelCase__ , keep_accents=UpperCAmelCase__ ) _a : Optional[int] = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(UpperCAmelCase__ , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) _a : Optional[int] = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( UpperCAmelCase__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) _a : int = tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) self.assertListEqual( UpperCAmelCase__ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) _a : Any = tokenizer.convert_ids_to_tokens(UpperCAmelCase__ ) self.assertListEqual( UpperCAmelCase__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """.""", ] , ) def _lowercase ( self : Any ) -> List[Any]: _a : Optional[Any] = (self.rust_tokenizer_class, """hf-internal-testing/tiny-random-nllb""", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _a : Optional[int] = self.rust_tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__ ) _a : Optional[Any] = self.tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__ ) _a : Union[str, Any] = tempfile.mkdtemp() _a : List[Any] = tokenizer_r.save_pretrained(UpperCAmelCase__ ) _a : List[Any] = tokenizer_p.save_pretrained(UpperCAmelCase__ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) _a : Union[str, Any] = tuple(f for f in tokenizer_r_files if """tokenizer.json""" not in f ) self.assertSequenceEqual(UpperCAmelCase__ , UpperCAmelCase__ ) # Checks everything loads correctly in the same way _a : Tuple = tokenizer_r.from_pretrained(UpperCAmelCase__ ) _a : Tuple = tokenizer_p.from_pretrained(UpperCAmelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCAmelCase__ , UpperCAmelCase__ ) ) shutil.rmtree(UpperCAmelCase__ ) # Save tokenizer rust, legacy_format=True _a : Optional[Any] = tempfile.mkdtemp() _a : Optional[Any] = tokenizer_r.save_pretrained(UpperCAmelCase__ , legacy_format=UpperCAmelCase__ ) _a : Dict = tokenizer_p.save_pretrained(UpperCAmelCase__ ) # Checks it save with the same files self.assertSequenceEqual(UpperCAmelCase__ , UpperCAmelCase__ ) # Checks everything loads correctly in the same way _a : int = tokenizer_r.from_pretrained(UpperCAmelCase__ ) _a : Any = tokenizer_p.from_pretrained(UpperCAmelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCAmelCase__ , UpperCAmelCase__ ) ) shutil.rmtree(UpperCAmelCase__ ) # Save tokenizer rust, legacy_format=False _a : Optional[Any] = tempfile.mkdtemp() _a : List[str] = tokenizer_r.save_pretrained(UpperCAmelCase__ , legacy_format=UpperCAmelCase__ ) _a : List[str] = tokenizer_p.save_pretrained(UpperCAmelCase__ ) # Checks it saved the tokenizer.json file self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way _a : Tuple = tokenizer_r.from_pretrained(UpperCAmelCase__ ) _a : Any = tokenizer_p.from_pretrained(UpperCAmelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCAmelCase__ , UpperCAmelCase__ ) ) shutil.rmtree(UpperCAmelCase__ ) @require_torch def _lowercase ( self : str ) -> List[str]: if not self.test_seqaseq: return _a : Tuple = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): # Longer text that will definitely require truncation. _a : Union[str, Any] = [ """ UN Chief Says There Is No Military Solution in Syria""", """ Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for""" """ Syria is that 'there is no military solution' to the nearly five-year conflict and more weapons""" """ will only worsen the violence and misery for millions of people.""", ] _a : Dict = [ """Şeful ONU declară că nu există o soluţie militară în Siria""", """Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al""" """ Rusiei pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi""" """ că noi arme nu vor face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.""", ] try: _a : Optional[int] = tokenizer.prepare_seqaseq_batch( src_texts=UpperCAmelCase__ , tgt_texts=UpperCAmelCase__ , max_length=3 , max_target_length=10 , return_tensors="""pt""" , src_lang="""eng_Latn""" , tgt_lang="""ron_Latn""" , ) except NotImplementedError: return self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.labels.shape[1] , 10 ) # max_target_length will default to max_length if not specified _a : List[str] = tokenizer.prepare_seqaseq_batch( UpperCAmelCase__ , tgt_texts=UpperCAmelCase__ , max_length=3 , return_tensors="""pt""" ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.labels.shape[1] , 3 ) _a : List[Any] = tokenizer.prepare_seqaseq_batch( src_texts=UpperCAmelCase__ , max_length=3 , max_target_length=10 , return_tensors="""pt""" ) self.assertEqual(batch_encoder_only.input_ids.shape[1] , 3 ) self.assertEqual(batch_encoder_only.attention_mask.shape[1] , 3 ) self.assertNotIn("""decoder_input_ids""" , UpperCAmelCase__ ) @unittest.skip("""Unfortunately way too slow to build a BPE with SentencePiece.""" ) def _lowercase ( self : List[Any] ) -> Dict: pass def _lowercase ( self : Any ) -> Optional[Any]: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _a : Tuple = [AddedToken("""<special>""" , lstrip=UpperCAmelCase__ )] _a : Union[str, Any] = self.rust_tokenizer_class.from_pretrained( UpperCAmelCase__ , additional_special_tokens=UpperCAmelCase__ , **UpperCAmelCase__ ) _a : Optional[int] = tokenizer_r.encode("""Hey this is a <special> token""" ) _a : Optional[Any] = tokenizer_r.encode("""<special>""" , add_special_tokens=UpperCAmelCase__ )[0] self.assertTrue(special_token_id in r_output ) if self.test_slow_tokenizer: _a : str = self.rust_tokenizer_class.from_pretrained( UpperCAmelCase__ , additional_special_tokens=UpperCAmelCase__ , **UpperCAmelCase__ , ) _a : Tuple = self.tokenizer_class.from_pretrained( UpperCAmelCase__ , additional_special_tokens=UpperCAmelCase__ , **UpperCAmelCase__ ) _a : Any = tokenizer_p.encode("""Hey this is a <special> token""" ) _a : Optional[int] = tokenizer_cr.encode("""Hey this is a <special> token""" ) self.assertEqual(UpperCAmelCase__ , UpperCAmelCase__ ) self.assertEqual(UpperCAmelCase__ , UpperCAmelCase__ ) self.assertTrue(special_token_id in p_output ) self.assertTrue(special_token_id in cr_output ) @require_torch @require_sentencepiece @require_tokenizers class UpperCamelCase ( unittest.TestCase ): UpperCamelCase : Optional[int] = '''facebook/nllb-200-distilled-600M''' UpperCamelCase : Any = [ ''' UN Chief Says There Is No Military Solution in Syria''', ''' Secretary-General Ban Ki-moon says his response to Russia\'s stepped up military support for Syria is that "there is no military solution" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.''', ] UpperCamelCase : str = [ '''Şeful ONU declară că nu există o soluţie militară în Siria''', '''Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei''' ''' pentru Siria este că "nu există o soluţie militară" la conflictul de aproape cinci ani şi că noi arme nu vor''' ''' face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.''', ] UpperCamelCase : int = [ 256_047, 16_297, 134_408, 8_165, 248_066, 14_734, 950, 1_135, 105_721, 3_573, 83, 27_352, 108, 49_486, 2, ] @classmethod def _lowercase ( cls : List[Any] ) -> List[Any]: _a : NllbTokenizer = NllbTokenizer.from_pretrained( cls.checkpoint_name , src_lang="""eng_Latn""" , tgt_lang="""ron_Latn""" ) _a : Union[str, Any] = 1 return cls def _lowercase ( self : Dict ) -> Any: self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""ace_Arab"""] , 256001 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""ace_Latn"""] , 256002 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""fra_Latn"""] , 256057 ) def _lowercase ( self : str ) -> str: _a : str = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , UpperCAmelCase__ ) def _lowercase ( self : List[str] ) -> List[Any]: self.assertIn(UpperCAmelCase__ , self.tokenizer.all_special_ids ) # fmt: off _a : int = [RO_CODE, 4254, 98068, 112923, 39072, 3909, 713, 102767, 26, 17314, 35642, 14683, 33118, 2022, 66987, 2, 256047] # fmt: on _a : int = self.tokenizer.decode(UpperCAmelCase__ , skip_special_tokens=UpperCAmelCase__ ) _a : Optional[int] = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=UpperCAmelCase__ ) self.assertEqual(UpperCAmelCase__ , UpperCAmelCase__ ) self.assertNotIn(self.tokenizer.eos_token , UpperCAmelCase__ ) def _lowercase ( self : Any ) -> List[str]: _a : Tuple = ["""this is gunna be a long sentence """ * 20] assert isinstance(src_text[0] , UpperCAmelCase__ ) _a : Tuple = 10 _a : int = self.tokenizer(UpperCAmelCase__ , max_length=UpperCAmelCase__ , truncation=UpperCAmelCase__ ).input_ids[0] self.assertEqual(ids[-1] , 2 ) self.assertEqual(ids[0] , UpperCAmelCase__ ) self.assertEqual(len(UpperCAmelCase__ ) , UpperCAmelCase__ ) def _lowercase ( self : Tuple ) -> Optional[Any]: self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["""<mask>""", """ar_AR"""] ) , [256203, 3] ) def _lowercase ( self : int ) -> int: _a : int = tempfile.mkdtemp() _a : Union[str, Any] = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(UpperCAmelCase__ ) _a : Tuple = NllbTokenizer.from_pretrained(UpperCAmelCase__ ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , UpperCAmelCase__ ) @require_torch def _lowercase ( self : List[Any] ) -> Tuple: _a : Optional[int] = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=UpperCAmelCase__ , truncation=UpperCAmelCase__ , max_length=len(self.expected_src_tokens ) , return_tensors="""pt""" , ) _a : Tuple = shift_tokens_right( batch["""labels"""] , self.tokenizer.pad_token_id , self.tokenizer.lang_code_to_id["""ron_Latn"""] ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) self.assertEqual((2, 15) , batch.input_ids.shape ) self.assertEqual((2, 15) , batch.attention_mask.shape ) _a : Optional[Any] = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , UpperCAmelCase__ ) self.assertEqual(UpperCAmelCase__ , batch.decoder_input_ids[0, 0] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [EN_CODE] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) def _lowercase ( self : Optional[Any] ) -> Optional[Any]: _a : Dict = self.tokenizer(self.src_text , padding=UpperCAmelCase__ , truncation=UpperCAmelCase__ , max_length=3 , return_tensors="""pt""" ) _a : List[str] = self.tokenizer( text_target=self.tgt_text , padding=UpperCAmelCase__ , truncation=UpperCAmelCase__ , max_length=10 , return_tensors="""pt""" ) _a : Optional[int] = targets["""input_ids"""] _a : List[Any] = shift_tokens_right( UpperCAmelCase__ , self.tokenizer.pad_token_id , decoder_start_token_id=self.tokenizer.lang_code_to_id[self.tokenizer.tgt_lang] , ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def _lowercase ( self : List[str] ) -> Any: _a : List[str] = self.tokenizer._build_translation_inputs( """A test""" , return_tensors="""pt""" , src_lang="""eng_Latn""" , tgt_lang="""fra_Latn""" ) self.assertEqual( nested_simplify(UpperCAmelCase__ ) , { # A, test, EOS, en_XX """input_ids""": [[256047, 70, 7356, 2]], """attention_mask""": [[1, 1, 1, 1]], # ar_AR """forced_bos_token_id""": 256057, } , ) @require_torch def _lowercase ( self : Optional[Any] ) -> Dict: _a : str = True _a : List[str] = self.tokenizer( """UN Chief says there is no military solution in Syria""" , src_lang="""eng_Latn""" , tgt_lang="""fra_Latn""" ) self.assertEqual( inputs.input_ids , [16297, 134408, 25653, 6370, 248, 254, 103929, 94995, 108, 49486, 2, 256047] ) _a : Any = False _a : int = self.tokenizer( """UN Chief says there is no military solution in Syria""" , src_lang="""eng_Latn""" , tgt_lang="""fra_Latn""" ) self.assertEqual( inputs.input_ids , [256047, 16297, 134408, 25653, 6370, 248, 254, 103929, 94995, 108, 49486, 2] )
294
"""simple docstring""" def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' # Check if the input is valid if not len(UpperCamelCase__ ) == len(UpperCamelCase__ ) == 3: raise ValueError("""Please enter a valid equation.""" ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError("""Both a & b of two equations can't be zero.""" ) # Extract the coefficients _a , _a , _a : Any = equationa _a , _a , _a : Tuple = equationa # Calculate the determinants of the matrices _a : int = aa * ba - aa * ba _a : str = ca * ba - ca * ba _a : str = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError("""Infinite solutions. (Consistent system)""" ) else: raise ValueError("""No solution. (Inconsistent system)""" ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: _a : Dict = determinant_x / determinant _a : str = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
294
1
"""simple docstring""" import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json', } class UpperCamelCase ( snake_case_ ): UpperCamelCase : int = '''mvp''' UpperCamelCase : Union[str, Any] = ['''past_key_values'''] UpperCamelCase : Any = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : List[str] , UpperCAmelCase__ : List[str]=50267 , UpperCAmelCase__ : Optional[Any]=1024 , UpperCAmelCase__ : Tuple=12 , UpperCAmelCase__ : Optional[Any]=4096 , UpperCAmelCase__ : int=16 , UpperCAmelCase__ : Tuple=12 , UpperCAmelCase__ : int=4096 , UpperCAmelCase__ : List[Any]=16 , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : Tuple="gelu" , UpperCAmelCase__ : Union[str, Any]=1024 , UpperCAmelCase__ : List[str]=0.1 , UpperCAmelCase__ : Any=0.0 , UpperCAmelCase__ : Dict=0.0 , UpperCAmelCase__ : Tuple=0.0_2 , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : Optional[Any]=False , UpperCAmelCase__ : int=True , UpperCAmelCase__ : Tuple=1 , UpperCAmelCase__ : Dict=0 , UpperCAmelCase__ : Union[str, Any]=2 , UpperCAmelCase__ : Optional[int]=True , UpperCAmelCase__ : Tuple=2 , UpperCAmelCase__ : Any=2 , UpperCAmelCase__ : Optional[Any]=False , UpperCAmelCase__ : Dict=100 , UpperCAmelCase__ : Union[str, Any]=800 , **UpperCAmelCase__ : Dict , ) -> List[Any]: _a : Any = vocab_size _a : Any = max_position_embeddings _a : Union[str, Any] = d_model _a : List[str] = encoder_ffn_dim _a : List[Any] = encoder_layers _a : Dict = encoder_attention_heads _a : Tuple = decoder_ffn_dim _a : List[Any] = decoder_layers _a : Optional[Any] = decoder_attention_heads _a : Optional[Any] = dropout _a : str = attention_dropout _a : Dict = activation_dropout _a : Any = activation_function _a : Tuple = init_std _a : Dict = encoder_layerdrop _a : Optional[int] = decoder_layerdrop _a : Optional[Any] = classifier_dropout _a : List[Any] = use_cache _a : Dict = encoder_layers _a : str = scale_embedding # scale factor will be sqrt(d_model) if True _a : int = use_prompt _a : Dict = prompt_length _a : Dict = prompt_mid_dim super().__init__( pad_token_id=UpperCAmelCase__ , bos_token_id=UpperCAmelCase__ , eos_token_id=UpperCAmelCase__ , is_encoder_decoder=UpperCAmelCase__ , decoder_start_token_id=UpperCAmelCase__ , forced_eos_token_id=UpperCAmelCase__ , **UpperCAmelCase__ , ) if self.forced_bos_token_id is None and kwargs.get("""force_bos_token_to_be_generated""" , UpperCAmelCase__ ): _a : List[str] = self.bos_token_id warnings.warn( f"""Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. """ """The config can simply be saved and uploaded again to be fixed.""" )
294
"""simple docstring""" _snake_case = '\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n' _snake_case = [{'type': 'code', 'content': INSTALL_CONTENT}] _snake_case = { '{processor_class}': 'FakeProcessorClass', '{model_class}': 'FakeModelClass', '{object_class}': 'FakeObjectClass', }
294
1
"""simple docstring""" import re from filelock import FileLock try: import nltk _snake_case = True except (ImportError, ModuleNotFoundError): _snake_case = False if NLTK_AVAILABLE: with FileLock('.lock') as lock: nltk.download('punkt', quiet=True) def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' re.sub("""<n>""" , """""" , UpperCamelCase__ ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(UpperCamelCase__ ) )
294
"""simple docstring""" import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json', } class UpperCamelCase ( snake_case_ ): UpperCamelCase : int = '''mvp''' UpperCamelCase : Union[str, Any] = ['''past_key_values'''] UpperCamelCase : Any = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : List[str] , UpperCAmelCase__ : List[str]=50267 , UpperCAmelCase__ : Optional[Any]=1024 , UpperCAmelCase__ : Tuple=12 , UpperCAmelCase__ : Optional[Any]=4096 , UpperCAmelCase__ : int=16 , UpperCAmelCase__ : Tuple=12 , UpperCAmelCase__ : int=4096 , UpperCAmelCase__ : List[Any]=16 , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : Tuple="gelu" , UpperCAmelCase__ : Union[str, Any]=1024 , UpperCAmelCase__ : List[str]=0.1 , UpperCAmelCase__ : Any=0.0 , UpperCAmelCase__ : Dict=0.0 , UpperCAmelCase__ : Tuple=0.0_2 , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : Optional[Any]=False , UpperCAmelCase__ : int=True , UpperCAmelCase__ : Tuple=1 , UpperCAmelCase__ : Dict=0 , UpperCAmelCase__ : Union[str, Any]=2 , UpperCAmelCase__ : Optional[int]=True , UpperCAmelCase__ : Tuple=2 , UpperCAmelCase__ : Any=2 , UpperCAmelCase__ : Optional[Any]=False , UpperCAmelCase__ : Dict=100 , UpperCAmelCase__ : Union[str, Any]=800 , **UpperCAmelCase__ : Dict , ) -> List[Any]: _a : Any = vocab_size _a : Any = max_position_embeddings _a : Union[str, Any] = d_model _a : List[str] = encoder_ffn_dim _a : List[Any] = encoder_layers _a : Dict = encoder_attention_heads _a : Tuple = decoder_ffn_dim _a : List[Any] = decoder_layers _a : Optional[Any] = decoder_attention_heads _a : Optional[Any] = dropout _a : str = attention_dropout _a : Dict = activation_dropout _a : Any = activation_function _a : Tuple = init_std _a : Dict = encoder_layerdrop _a : Optional[int] = decoder_layerdrop _a : Optional[Any] = classifier_dropout _a : List[Any] = use_cache _a : Dict = encoder_layers _a : str = scale_embedding # scale factor will be sqrt(d_model) if True _a : int = use_prompt _a : Dict = prompt_length _a : Dict = prompt_mid_dim super().__init__( pad_token_id=UpperCAmelCase__ , bos_token_id=UpperCAmelCase__ , eos_token_id=UpperCAmelCase__ , is_encoder_decoder=UpperCAmelCase__ , decoder_start_token_id=UpperCAmelCase__ , forced_eos_token_id=UpperCAmelCase__ , **UpperCAmelCase__ , ) if self.forced_bos_token_id is None and kwargs.get("""force_bos_token_to_be_generated""" , UpperCAmelCase__ ): _a : List[str] = self.bos_token_id warnings.warn( f"""Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. """ """The config can simply be saved and uploaded again to be fixed.""" )
294
1
"""simple docstring""" import tempfile import unittest import numpy as np from diffusers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionPipeline, PNDMScheduler, ) from diffusers.utils.testing_utils import is_onnx_available, nightly, require_onnxruntime, require_torch_gpu from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class UpperCamelCase ( snake_case_ , unittest.TestCase ): UpperCamelCase : List[str] = '''hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline''' def _lowercase ( self : Tuple , UpperCAmelCase__ : Union[str, Any]=0 ) -> int: _a : Dict = np.random.RandomState(UpperCAmelCase__ ) _a : str = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 7.5, """output_type""": """numpy""", } return inputs def _lowercase ( self : Any ) -> Union[str, Any]: _a : Any = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Any = self.get_dummy_inputs() _a : Union[str, Any] = pipe(**UpperCAmelCase__ ).images _a : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _a : Optional[int] = np.array([0.6_5_0_7_2, 0.5_8_4_9_2, 0.4_8_2_1_9, 0.5_5_5_2_1, 0.5_3_1_8_0, 0.5_5_9_3_9, 0.5_0_6_9_7, 0.3_9_8_0_0, 0.4_6_4_5_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def _lowercase ( self : Union[str, Any] ) -> int: _a : Optional[Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) _a : Dict = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : List[str] = self.get_dummy_inputs() _a : List[str] = pipe(**UpperCAmelCase__ ).images _a : Union[str, Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _a : str = np.array([0.6_5_8_6_3, 0.5_9_4_2_5, 0.4_9_3_2_6, 0.5_6_3_1_3, 0.5_3_8_7_5, 0.5_6_6_2_7, 0.5_1_0_6_5, 0.3_9_7_7_7, 0.4_6_3_3_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def _lowercase ( self : int ) -> Any: _a : Union[str, Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) _a : Tuple = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Dict = self.get_dummy_inputs() _a : Any = pipe(**UpperCAmelCase__ ).images _a : Any = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _a : Union[str, Any] = np.array([0.5_3_7_5_5, 0.6_0_7_8_6, 0.4_7_4_0_2, 0.4_9_4_8_8, 0.5_1_8_6_9, 0.4_9_8_1_9, 0.4_7_9_8_5, 0.3_8_9_5_7, 0.4_4_2_7_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def _lowercase ( self : Optional[int] ) -> str: _a : Tuple = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) _a : Union[str, Any] = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Optional[Any] = self.get_dummy_inputs() _a : int = pipe(**UpperCAmelCase__ ).images _a : Any = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _a : List[Any] = np.array([0.5_3_7_5_5, 0.6_0_7_8_6, 0.4_7_4_0_2, 0.4_9_4_8_8, 0.5_1_8_6_9, 0.4_9_8_1_9, 0.4_7_9_8_5, 0.3_8_9_5_7, 0.4_4_2_7_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def _lowercase ( self : Optional[int] ) -> int: _a : Tuple = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) _a : Union[str, Any] = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : List[Any] = self.get_dummy_inputs() _a : Union[str, Any] = pipe(**UpperCAmelCase__ ).images _a : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _a : str = np.array([0.5_3_8_1_7, 0.6_0_8_1_2, 0.4_7_3_8_4, 0.4_9_5_3_0, 0.5_1_8_9_4, 0.4_9_8_1_4, 0.4_7_9_8_4, 0.3_8_9_5_8, 0.4_4_2_7_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def _lowercase ( self : List[str] ) -> str: _a : Dict = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) _a : Optional[Any] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Union[str, Any] = self.get_dummy_inputs() _a : Any = pipe(**UpperCAmelCase__ ).images _a : int = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) _a : Optional[int] = np.array([0.5_3_8_9_5, 0.6_0_8_0_8, 0.4_7_9_3_3, 0.4_9_6_0_8, 0.5_1_8_8_6, 0.4_9_9_5_0, 0.4_8_0_5_3, 0.3_8_9_5_7, 0.4_4_2_0_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def _lowercase ( self : List[str] ) -> List[str]: _a : str = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : List[Any] = self.get_dummy_inputs() _a : Optional[int] = 3 * [inputs["""prompt"""]] # forward _a : Dict = pipe(**UpperCAmelCase__ ) _a : Tuple = output.images[0, -3:, -3:, -1] _a : str = self.get_dummy_inputs() _a : Union[str, Any] = 3 * [inputs.pop("""prompt""" )] _a : Any = pipe.tokenizer( UpperCAmelCase__ , padding="""max_length""" , max_length=pipe.tokenizer.model_max_length , truncation=UpperCAmelCase__ , return_tensors="""np""" , ) _a : Tuple = text_inputs["""input_ids"""] _a : Dict = pipe.text_encoder(input_ids=text_inputs.astype(np.intaa ) )[0] _a : int = prompt_embeds # forward _a : List[str] = pipe(**UpperCAmelCase__ ) _a : Tuple = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1E-4 def _lowercase ( self : Any ) -> int: _a : Optional[Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Tuple = self.get_dummy_inputs() _a : List[str] = 3 * ["""this is a negative prompt"""] _a : Optional[Any] = negative_prompt _a : Optional[int] = 3 * [inputs["""prompt"""]] # forward _a : Any = pipe(**UpperCAmelCase__ ) _a : List[Any] = output.images[0, -3:, -3:, -1] _a : int = self.get_dummy_inputs() _a : int = 3 * [inputs.pop("""prompt""" )] _a : Tuple = [] for p in [prompt, negative_prompt]: _a : Optional[int] = pipe.tokenizer( UpperCAmelCase__ , padding="""max_length""" , max_length=pipe.tokenizer.model_max_length , truncation=UpperCAmelCase__ , return_tensors="""np""" , ) _a : Optional[int] = text_inputs["""input_ids"""] embeds.append(pipe.text_encoder(input_ids=text_inputs.astype(np.intaa ) )[0] ) _a , _a : str = embeds # forward _a : List[str] = pipe(**UpperCAmelCase__ ) _a : Optional[int] = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1E-4 @nightly @require_onnxruntime @require_torch_gpu class UpperCamelCase ( unittest.TestCase ): @property def _lowercase ( self : int ) -> Optional[int]: return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def _lowercase ( self : Optional[Any] ) -> List[Any]: _a : Optional[Any] = ort.SessionOptions() _a : Dict = False return options def _lowercase ( self : str ) -> Any: # using the PNDM scheduler by default _a : int = OnnxStableDiffusionPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""onnx""" , safety_checker=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Optional[Any] = """A painting of a squirrel eating a burger""" np.random.seed(0 ) _a : Dict = sd_pipe([prompt] , guidance_scale=6.0 , num_inference_steps=10 , output_type="""np""" ) _a : Union[str, Any] = output.images _a : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _a : Optional[Any] = np.array([0.0_4_5_2, 0.0_3_9_0, 0.0_0_8_7, 0.0_3_5_0, 0.0_6_1_7, 0.0_3_6_4, 0.0_5_4_4, 0.0_5_2_3, 0.0_7_2_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowercase ( self : Union[str, Any] ) -> List[str]: _a : Any = DDIMScheduler.from_pretrained( """runwayml/stable-diffusion-v1-5""" , subfolder="""scheduler""" , revision="""onnx""" ) _a : str = OnnxStableDiffusionPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , revision="""onnx""" , scheduler=UpperCAmelCase__ , safety_checker=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Tuple = """open neural network exchange""" _a : List[str] = np.random.RandomState(0 ) _a : Optional[int] = sd_pipe([prompt] , guidance_scale=7.5 , num_inference_steps=10 , generator=UpperCAmelCase__ , output_type="""np""" ) _a : List[Any] = output.images _a : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _a : Dict = np.array([0.2_8_6_7, 0.1_9_7_4, 0.1_4_8_1, 0.7_2_9_4, 0.7_2_5_1, 0.6_6_6_7, 0.4_1_9_4, 0.5_6_4_2, 0.6_4_8_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowercase ( self : Optional[Any] ) -> List[Any]: _a : Union[str, Any] = LMSDiscreteScheduler.from_pretrained( """runwayml/stable-diffusion-v1-5""" , subfolder="""scheduler""" , revision="""onnx""" ) _a : Optional[Any] = OnnxStableDiffusionPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , revision="""onnx""" , scheduler=UpperCAmelCase__ , safety_checker=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Any = """open neural network exchange""" _a : List[str] = np.random.RandomState(0 ) _a : Optional[Any] = sd_pipe([prompt] , guidance_scale=7.5 , num_inference_steps=10 , generator=UpperCAmelCase__ , output_type="""np""" ) _a : List[str] = output.images _a : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _a : List[str] = np.array([0.2_3_0_6, 0.1_9_5_9, 0.1_5_9_3, 0.6_5_4_9, 0.6_3_9_4, 0.5_4_0_8, 0.5_0_6_5, 0.6_0_1_0, 0.6_1_6_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _lowercase ( self : Optional[int] ) -> str: _a : Optional[int] = 0 def test_callback_fn(UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : np.ndarray ) -> None: _a : Dict = True nonlocal number_of_steps number_of_steps += 1 if step == 0: assert latents.shape == (1, 4, 64, 64) _a : str = latents[0, -3:, -3:, -1] _a : List[Any] = np.array( [-0.6_7_7_2, -0.3_8_3_5, -1.2_4_5_6, 0.1_9_0_5, -1.0_9_7_4, 0.6_9_6_7, -1.9_3_5_3, 0.0_1_7_8, 1.0_1_6_7] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 1E-3 elif step == 5: assert latents.shape == (1, 4, 64, 64) _a : Optional[Any] = latents[0, -3:, -3:, -1] _a : List[Any] = np.array( [-0.3_3_5_1, 0.2_2_4_1, -0.1_8_3_7, -0.2_3_2_5, -0.6_5_7_7, 0.3_3_9_3, -0.0_2_4_1, 0.5_8_9_9, 1.3_8_7_5] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 1E-3 _a : List[str] = False _a : Tuple = OnnxStableDiffusionPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , revision="""onnx""" , safety_checker=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Any = """Andromeda galaxy in a bottle""" _a : Dict = np.random.RandomState(0 ) pipe( prompt=UpperCAmelCase__ , num_inference_steps=5 , guidance_scale=7.5 , generator=UpperCAmelCase__ , callback=UpperCAmelCase__ , callback_steps=1 , ) assert test_callback_fn.has_been_called assert number_of_steps == 6 def _lowercase ( self : List[str] ) -> Union[str, Any]: _a : List[str] = OnnxStableDiffusionPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , revision="""onnx""" , safety_checker=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) assert isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) assert pipe.safety_checker is None _a : List[Any] = pipe("""example prompt""" , num_inference_steps=2 ).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(UpperCAmelCase__ ) _a : Optional[int] = OnnxStableDiffusionPipeline.from_pretrained(UpperCAmelCase__ ) # sanity check that the pipeline still works assert pipe.safety_checker is None _a : List[Any] = pipe("""example prompt""" , num_inference_steps=2 ).images[0] assert image is not None
294
"""simple docstring""" import logging import os from typing import List, TextIO, Union from conllu import parse_incr from utils_ner import InputExample, Split, TokenClassificationTask _snake_case = logging.getLogger(__name__) class UpperCamelCase ( snake_case_ ): def __init__( self : Optional[Any] , UpperCAmelCase__ : Optional[int]=-1 ) -> Tuple: # in NER datasets, the last column is usually reserved for NER label _a : Optional[int] = label_idx def _lowercase ( self : Any , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Union[Split, str] ) -> List[InputExample]: if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): _a : Any = mode.value _a : Optional[int] = os.path.join(UpperCAmelCase__ , f"""{mode}.txt""" ) _a : int = 1 _a : int = [] with open(UpperCAmelCase__ , encoding="""utf-8""" ) as f: _a : str = [] _a : str = [] for line in f: if line.startswith("""-DOCSTART-""" ) or line == "" or line == "\n": if words: examples.append(InputExample(guid=f"""{mode}-{guid_index}""" , words=UpperCAmelCase__ , labels=UpperCAmelCase__ ) ) guid_index += 1 _a : List[str] = [] _a : str = [] else: _a : List[Any] = line.split(""" """ ) words.append(splits[0] ) if len(UpperCAmelCase__ ) > 1: labels.append(splits[self.label_idx].replace("""\n""" , """""" ) ) else: # Examples could have no label for mode = "test" labels.append("""O""" ) if words: examples.append(InputExample(guid=f"""{mode}-{guid_index}""" , words=UpperCAmelCase__ , labels=UpperCAmelCase__ ) ) return examples def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : List ) -> Union[str, Any]: _a : List[str] = 0 for line in test_input_reader: if line.startswith("""-DOCSTART-""" ) or line == "" or line == "\n": writer.write(UpperCAmelCase__ ) if not preds_list[example_id]: example_id += 1 elif preds_list[example_id]: _a : int = line.split()[0] + """ """ + preds_list[example_id].pop(0 ) + """\n""" writer.write(UpperCAmelCase__ ) else: logger.warning("""Maximum sequence length exceeded: No prediction for '%s'.""" , line.split()[0] ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : str ) -> List[str]: if path: with open(UpperCAmelCase__ , """r""" ) as f: _a : List[Any] = f.read().splitlines() if "O" not in labels: _a : Union[str, Any] = ["""O"""] + labels return labels else: return ["O", "B-MISC", "I-MISC", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC"] class UpperCamelCase ( snake_case_ ): def __init__( self : Union[str, Any] ) -> List[str]: # in CONLL2003 dataset chunk column is second-to-last super().__init__(label_idx=-2 ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : str ) -> List[str]: if path: with open(UpperCAmelCase__ , """r""" ) as f: _a : Optional[int] = f.read().splitlines() if "O" not in labels: _a : Optional[Any] = ["""O"""] + labels return labels else: return [ "O", "B-ADVP", "B-INTJ", "B-LST", "B-PRT", "B-NP", "B-SBAR", "B-VP", "B-ADJP", "B-CONJP", "B-PP", "I-ADVP", "I-INTJ", "I-LST", "I-PRT", "I-NP", "I-SBAR", "I-VP", "I-ADJP", "I-CONJP", "I-PP", ] class UpperCamelCase ( snake_case_ ): def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Union[Split, str] ) -> List[InputExample]: if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): _a : List[Any] = mode.value _a : Union[str, Any] = os.path.join(UpperCAmelCase__ , f"""{mode}.txt""" ) _a : List[str] = 1 _a : Optional[Any] = [] with open(UpperCAmelCase__ , encoding="""utf-8""" ) as f: for sentence in parse_incr(UpperCAmelCase__ ): _a : List[Any] = [] _a : Any = [] for token in sentence: words.append(token["""form"""] ) labels.append(token["""upos"""] ) assert len(UpperCAmelCase__ ) == len(UpperCAmelCase__ ) if words: examples.append(InputExample(guid=f"""{mode}-{guid_index}""" , words=UpperCAmelCase__ , labels=UpperCAmelCase__ ) ) guid_index += 1 return examples def _lowercase ( self : Tuple , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : List ) -> Dict: _a : Optional[Any] = 0 for sentence in parse_incr(UpperCAmelCase__ ): _a : List[str] = preds_list[example_id] _a : str = """""" for token in sentence: out += f"""{token['form']} ({token['upos']}|{s_p.pop(0 )}) """ out += "\n" writer.write(UpperCAmelCase__ ) example_id += 1 def _lowercase ( self : List[str] , UpperCAmelCase__ : str ) -> List[str]: if path: with open(UpperCAmelCase__ , """r""" ) as f: return f.read().splitlines() else: return [ "ADJ", "ADP", "ADV", "AUX", "CCONJ", "DET", "INTJ", "NOUN", "NUM", "PART", "PRON", "PROPN", "PUNCT", "SCONJ", "SYM", "VERB", "X", ]
294
1
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.activations import gelu_new, gelu_python, get_activation @require_torch class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : List[str] ) -> List[Any]: _a : Union[str, Any] = torch.tensor([-100, -1, -0.1, 0, 0.1, 1.0, 100] ) _a : List[Any] = get_activation("""gelu""" ) self.assertTrue(torch.allclose(gelu_python(UpperCAmelCase__ ) , torch_builtin(UpperCAmelCase__ ) ) ) self.assertFalse(torch.allclose(gelu_python(UpperCAmelCase__ ) , gelu_new(UpperCAmelCase__ ) ) ) def _lowercase ( self : Optional[int] ) -> List[str]: _a : Optional[Any] = torch.tensor([-100, -1, -0.1, 0, 0.1, 1.0, 100] ) _a : Tuple = get_activation("""gelu""" ) _a : Any = get_activation("""gelu_10""" ) _a : Any = torch_builtin(UpperCAmelCase__ ) _a : Union[str, Any] = geluaa(UpperCAmelCase__ ) _a : Dict = torch.where(y_gelu_aa < 1_0.0 , 1 , 0 ) self.assertTrue(torch.max(UpperCAmelCase__ ).item() == 1_0.0 ) self.assertTrue(torch.allclose(y_gelu * clipped_mask , y_gelu_aa * clipped_mask ) ) def _lowercase ( self : int ) -> Optional[Any]: get_activation("""gelu""" ) get_activation("""gelu_10""" ) get_activation("""gelu_fast""" ) get_activation("""gelu_new""" ) get_activation("""gelu_python""" ) get_activation("""gelu_pytorch_tanh""" ) get_activation("""linear""" ) get_activation("""mish""" ) get_activation("""quick_gelu""" ) get_activation("""relu""" ) get_activation("""sigmoid""" ) get_activation("""silu""" ) get_activation("""swish""" ) get_activation("""tanh""" ) with self.assertRaises(UpperCAmelCase__ ): get_activation("""bogus""" ) with self.assertRaises(UpperCAmelCase__ ): get_activation(UpperCAmelCase__ ) def _lowercase ( self : Dict ) -> Tuple: _a : int = get_activation("""gelu""" ) _a : str = 1 _a : Optional[int] = get_activation("""gelu""" ) self.assertEqual(acta.a , 1 ) with self.assertRaises(UpperCAmelCase__ ): _a : List[Any] = acta.a
294
"""simple docstring""" from __future__ import annotations import time import numpy as np _snake_case = [8, 5, 9, 7] _snake_case = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] _snake_case = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class UpperCamelCase : def __init__( self : List[Any] , UpperCAmelCase__ : list[int] , UpperCAmelCase__ : list[list[int]] , UpperCAmelCase__ : list[list[int]] , ) -> None: _a : List[str] = claim_vector _a : List[Any] = allocated_resources_table _a : Union[str, Any] = maximum_claim_table def _lowercase ( self : Tuple ) -> list[int]: return [ sum(p_item[i] for p_item in self.__allocated_resources_table ) for i in range(len(self.__allocated_resources_table[0] ) ) ] def _lowercase ( self : int ) -> list[int]: return np.array(self.__claim_vector ) - np.array( self.__processes_resource_summation() ) def _lowercase ( self : List[str] ) -> list[list[int]]: return [ list(np.array(self.__maximum_claim_table[i] ) - np.array(UpperCAmelCase__ ) ) for i, allocated_resource in enumerate(self.__allocated_resources_table ) ] def _lowercase ( self : Optional[Any] ) -> dict[int, list[int]]: return {self.__need().index(UpperCAmelCase__ ): i for i in self.__need()} def _lowercase ( self : Dict , **UpperCAmelCase__ : Optional[Any] ) -> None: _a : List[Any] = self.__need() _a : Optional[int] = self.__allocated_resources_table _a : str = self.__available_resources() _a : Optional[Any] = self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print("""_""" * 50 + """\n""" ) while need_list: _a : int = False for each_need in need_list: _a : Optional[int] = True for index, need in enumerate(UpperCAmelCase__ ): if need > available_resources[index]: _a : List[Any] = False break if execution: _a : str = True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: _a : Any = original_need_index print(f"""Process {process_number + 1} is executing.""" ) # remove the process run from stack need_list.remove(UpperCAmelCase__ ) # update available/freed resources stack _a : Union[str, Any] = np.array(UpperCAmelCase__ ) + np.array( alloc_resources_table[process_number] ) print( """Updated available resource stack for processes: """ + """ """.join([str(UpperCAmelCase__ ) for x in available_resources] ) ) break if safe: print("""The process is in a safe state.\n""" ) else: print("""System in unsafe state. Aborting...\n""" ) break def _lowercase ( self : Any ) -> Optional[int]: print(""" """ * 9 + """Allocated Resource Table""" ) for item in self.__allocated_resources_table: print( f"""P{self.__allocated_resources_table.index(UpperCAmelCase__ ) + 1}""" + """ """.join(f"""{it:>8}""" for it in item ) + """\n""" ) print(""" """ * 9 + """System Resource Table""" ) for item in self.__maximum_claim_table: print( f"""P{self.__maximum_claim_table.index(UpperCAmelCase__ ) + 1}""" + """ """.join(f"""{it:>8}""" for it in item ) + """\n""" ) print( """Current Usage by Active Processes: """ + """ """.join(str(UpperCAmelCase__ ) for x in self.__claim_vector ) ) print( """Initial Available Resources: """ + """ """.join(str(UpperCAmelCase__ ) for x in self.__available_resources() ) ) time.sleep(1 ) if __name__ == "__main__": import doctest doctest.testmod()
294
1
"""simple docstring""" import itertools import os from collections import Counter, defaultdict from concurrent.futures import ThreadPoolExecutor, as_completed import numpy as np import datasets from .execute import check_correctness _snake_case = '\\n@misc{chen2021evaluating,\n title={Evaluating Large Language Models Trained on Code},\n author={Mark Chen and Jerry Tworek and Heewoo Jun and Qiming Yuan \\nand Henrique Ponde de Oliveira Pinto and Jared Kaplan and Harri Edwards \\nand Yuri Burda and Nicholas Joseph and Greg Brockman and Alex Ray \\nand Raul Puri and Gretchen Krueger and Michael Petrov and Heidy Khlaaf \\nand Girish Sastry and Pamela Mishkin and Brooke Chan and Scott Gray \\nand Nick Ryder and Mikhail Pavlov and Alethea Power and Lukasz Kaiser \\nand Mohammad Bavarian and Clemens Winter and Philippe Tillet \\nand Felipe Petroski Such and Dave Cummings and Matthias Plappert \\nand Fotios Chantzis and Elizabeth Barnes and Ariel Herbert-Voss \\nand William Hebgen Guss and Alex Nichol and Alex Paino and Nikolas Tezak \\nand Jie Tang and Igor Babuschkin and Suchir Balaji and Shantanu Jain \\nand William Saunders and Christopher Hesse and Andrew N. Carr \\nand Jan Leike and Josh Achiam and Vedant Misra and Evan Morikawa \\nand Alec Radford and Matthew Knight and Miles Brundage and Mira Murati \\nand Katie Mayer and Peter Welinder and Bob McGrew and Dario Amodei \\nand Sam McCandlish and Ilya Sutskever and Wojciech Zaremba},\n year={2021},\n eprint={2107.03374},\n archivePrefix={arXiv},\n primaryClass={cs.LG}\n}\n' _snake_case = '\\nThis metric implements the evaluation harness for the HumanEval problem solving dataset\ndescribed in the paper "Evaluating Large Language Models Trained on Code"\n(https://arxiv.org/abs/2107.03374).\n' _snake_case = '\nCalculates how good are predictions given some references, using certain scores\nArgs:\n predictions: list of candidates to evaluate. Each candidates should be a list\n of strings with several code candidates to solve the problem.\n references: a list with a test for each prediction. Each test should evaluate the\n correctness of a code candidate.\n k: number of code candidates to consider in the evaluation (Default: [1, 10, 100])\n num_workers: number of workers used to evaluate the canidate programs (Default: 4).\n timeout:\nReturns:\n pass_at_k: dict with pass rates for each k\n results: dict with granular results of each unittest\nExamples:\n >>> code_eval = datasets.load_metric("code_eval")\n >>> test_cases = ["assert add(2,3)==5"]\n >>> candidates = [["def add(a,b): return a*b", "def add(a, b): return a+b"]]\n >>> pass_at_k, results = code_eval.compute(references=test_cases, predictions=candidates, k=[1, 2])\n >>> print(pass_at_k)\n {\'pass@1\': 0.5, \'pass@2\': 1.0}\n' _snake_case = '\n################################################################################\n !!!WARNING!!!\n################################################################################\nThe "code_eval" metric executes untrusted model-generated code in Python.\nAlthough it is highly unlikely that model-generated code will do something\novertly malicious in response to this test suite, model-generated code may act\ndestructively due to a lack of model capability or alignment.\nUsers are strongly encouraged to sandbox this evaluation suite so that it\ndoes not perform destructive actions on their host or network. For more\ninformation on how OpenAI sandboxes its code, see the paper "Evaluating Large\nLanguage Models Trained on Code" (https://arxiv.org/abs/2107.03374).\n\nOnce you have read this disclaimer and taken appropriate precautions,\nset the environment variable HF_ALLOW_CODE_EVAL="1". Within Python you can to this\nwith:\n\n>>> import os\n>>> os.environ["HF_ALLOW_CODE_EVAL"] = "1"\n\n################################################################################\\n' _snake_case = 'The MIT License\n\nCopyright (c) OpenAI (https://openai.com)\n\nPermission is hereby granted, free of charge, to any person obtaining a copy\nof this software and associated documentation files (the "Software"), to deal\nin the Software without restriction, including without limitation the rights\nto use, copy, modify, merge, publish, distribute, sublicense, and/or sell\ncopies of the Software, and to permit persons to whom the Software is\nfurnished to do so, subject to the following conditions:\n\nThe above copyright notice and this permission notice shall be included in\nall copies or substantial portions of the Software.\n\nTHE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\nIMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\nFITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\nAUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\nLIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\nOUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN\nTHE SOFTWARE.' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCamelCase ( datasets.Metric ): def _lowercase ( self : Union[str, Any] ) -> Union[str, Any]: return datasets.MetricInfo( # This is the description that will appear on the metrics page. description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""string""" ) ), """references""": datasets.Value("""string""" ), } ) , homepage="""https://github.com/openai/human-eval""" , codebase_urls=["""https://github.com/openai/human-eval"""] , reference_urls=["""https://github.com/openai/human-eval"""] , license=_LICENSE , ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : str=[1, 10, 100] , UpperCAmelCase__ : Tuple=4 , UpperCAmelCase__ : Union[str, Any]=3.0 ) -> int: if os.getenv("""HF_ALLOW_CODE_EVAL""" , 0 ) != "1": raise ValueError(_WARNING ) if os.name == "nt": raise NotImplementedError("""This metric is currently not supported on Windows.""" ) with ThreadPoolExecutor(max_workers=UpperCAmelCase__ ) as executor: _a : Any = [] _a : Any = Counter() _a : Dict = 0 _a : Any = defaultdict(UpperCAmelCase__ ) for task_id, (candidates, test_case) in enumerate(zip(UpperCAmelCase__ , UpperCAmelCase__ ) ): for candidate in candidates: _a : int = candidate + """\n""" + test_case _a : Dict = (test_program, timeout, task_id, completion_id[task_id]) _a : Optional[Any] = executor.submit(UpperCAmelCase__ , *UpperCAmelCase__ ) futures.append(UpperCAmelCase__ ) completion_id[task_id] += 1 n_samples += 1 for future in as_completed(UpperCAmelCase__ ): _a : Optional[int] = future.result() results[result["task_id"]].append((result["""completion_id"""], result) ) _a , _a : Union[str, Any] = [], [] for result in results.values(): result.sort() _a : List[str] = [r[1]["""passed"""] for r in result] total.append(len(UpperCAmelCase__ ) ) correct.append(sum(UpperCAmelCase__ ) ) _a : Union[str, Any] = np.array(UpperCAmelCase__ ) _a : str = np.array(UpperCAmelCase__ ) _a : Any = k _a : Any = {f"""pass@{k}""": estimate_pass_at_k(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ).mean() for k in ks if (total >= k).all()} return pass_at_k, results def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' def estimator(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> float: if n - c < k: return 1.0 return 1.0 - np.prod(1.0 - k / np.arange(n - c + 1 , n + 1 ) ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ): _a : Union[str, Any] = itertools.repeat(UpperCamelCase__ , len(UpperCamelCase__ ) ) else: assert len(UpperCamelCase__ ) == len(UpperCamelCase__ ) _a : Optional[Any] = iter(UpperCamelCase__ ) return np.array([estimator(int(UpperCamelCase__ ) , int(UpperCamelCase__ ) , UpperCamelCase__ ) for n, c in zip(UpperCamelCase__ , UpperCamelCase__ )] )
294
"""simple docstring""" from collections.abc import Iterable from typing import Generic, TypeVar _snake_case = TypeVar('_T') class UpperCamelCase ( Generic[_T] ): def __init__( self : Optional[int] , UpperCAmelCase__ : Iterable[_T] | None = None ) -> None: _a : list[_T] = list(iterable or [] ) _a : list[_T] = [] def __len__( self : str ) -> int: return len(self._stacka ) + len(self._stacka ) def __repr__( self : List[str] ) -> str: return f"""Queue({tuple(self._stacka[::-1] + self._stacka )})""" def _lowercase ( self : Union[str, Any] , UpperCAmelCase__ : _T ) -> None: self._stacka.append(UpperCAmelCase__ ) def _lowercase ( self : Optional[Any] ) -> _T: _a : Any = self._stacka.pop _a : Union[str, Any] = self._stacka.append if not self._stacka: while self._stacka: stacka_append(stacka_pop() ) if not self._stacka: raise IndexError("""Queue is empty""" ) return self._stacka.pop() if __name__ == "__main__": from doctest import testmod testmod()
294
1
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import AutoImageProcessor, SwinvaConfig, SwinvaForImageClassification def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : Any = SwinvaConfig() _a : Optional[int] = swinva_name.split("""_""" ) _a : Tuple = name_split[1] if "to" in name_split[3]: _a : Union[str, Any] = int(name_split[3][-3:] ) else: _a : int = int(name_split[3] ) if "to" in name_split[2]: _a : Dict = int(name_split[2][-2:] ) else: _a : Optional[int] = int(name_split[2][6:] ) if model_size == "tiny": _a : Any = 9_6 _a : Optional[int] = (2, 2, 6, 2) _a : Dict = (3, 6, 1_2, 2_4) elif model_size == "small": _a : Optional[int] = 9_6 _a : List[Any] = (2, 2, 1_8, 2) _a : int = (3, 6, 1_2, 2_4) elif model_size == "base": _a : str = 1_2_8 _a : int = (2, 2, 1_8, 2) _a : Any = (4, 8, 1_6, 3_2) else: _a : Tuple = 1_9_2 _a : List[Any] = (2, 2, 1_8, 2) _a : Union[str, Any] = (6, 1_2, 2_4, 4_8) if "to" in swinva_name: _a : Optional[int] = (1_2, 1_2, 1_2, 6) if ("22k" in swinva_name) and ("to" not in swinva_name): _a : str = 2_1_8_4_1 _a : List[Any] = """huggingface/label-files""" _a : List[str] = """imagenet-22k-id2label.json""" _a : Union[str, Any] = json.load(open(hf_hub_download(UpperCamelCase__ , UpperCamelCase__ , repo_type="""dataset""" ) , """r""" ) ) _a : str = {int(UpperCamelCase__ ): v for k, v in idalabel.items()} _a : int = idalabel _a : Optional[int] = {v: k for k, v in idalabel.items()} else: _a : str = 1_0_0_0 _a : int = """huggingface/label-files""" _a : Optional[int] = """imagenet-1k-id2label.json""" _a : str = json.load(open(hf_hub_download(UpperCamelCase__ , UpperCamelCase__ , repo_type="""dataset""" ) , """r""" ) ) _a : Tuple = {int(UpperCamelCase__ ): v for k, v in idalabel.items()} _a : Any = idalabel _a : int = {v: k for k, v in idalabel.items()} _a : List[Any] = img_size _a : List[Any] = num_classes _a : List[str] = embed_dim _a : Dict = depths _a : List[str] = num_heads _a : Optional[Any] = window_size return config def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' if "patch_embed.proj" in name: _a : str = name.replace("""patch_embed.proj""" , """embeddings.patch_embeddings.projection""" ) if "patch_embed.norm" in name: _a : Union[str, Any] = name.replace("""patch_embed.norm""" , """embeddings.norm""" ) if "layers" in name: _a : List[Any] = """encoder.""" + name if "attn.proj" in name: _a : Union[str, Any] = name.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in name: _a : List[str] = name.replace("""attn""" , """attention.self""" ) if "norm1" in name: _a : str = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: _a : List[Any] = name.replace("""norm2""" , """layernorm_after""" ) if "mlp.fc1" in name: _a : Any = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: _a : Dict = name.replace("""mlp.fc2""" , """output.dense""" ) if "q_bias" in name: _a : Tuple = name.replace("""q_bias""" , """query.bias""" ) if "k_bias" in name: _a : str = name.replace("""k_bias""" , """key.bias""" ) if "v_bias" in name: _a : int = name.replace("""v_bias""" , """value.bias""" ) if "cpb_mlp" in name: _a : Union[str, Any] = name.replace("""cpb_mlp""" , """continuous_position_bias_mlp""" ) if name == "norm.weight": _a : int = """layernorm.weight""" if name == "norm.bias": _a : List[Any] = """layernorm.bias""" if "head" in name: _a : Dict = name.replace("""head""" , """classifier""" ) else: _a : Union[str, Any] = """swinv2.""" + name return name def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' for key in orig_state_dict.copy().keys(): _a : Optional[Any] = orig_state_dict.pop(UpperCamelCase__ ) if "mask" in key: continue elif "qkv" in key: _a : str = key.split(""".""" ) _a : Union[str, Any] = int(key_split[1] ) _a : Tuple = int(key_split[3] ) _a : List[str] = model.swinva.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: _a : Union[str, Any] = val[:dim, :] _a : str = val[dim : dim * 2, :] _a : Union[str, Any] = val[-dim:, :] else: _a : Union[str, Any] = val[:dim] _a : Union[str, Any] = val[ dim : dim * 2 ] _a : int = val[-dim:] else: _a : Union[str, Any] = val return orig_state_dict def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : int = timm.create_model(UpperCamelCase__ , pretrained=UpperCamelCase__ ) timm_model.eval() _a : Union[str, Any] = get_swinva_config(UpperCamelCase__ ) _a : Tuple = SwinvaForImageClassification(UpperCamelCase__ ) model.eval() _a : int = convert_state_dict(timm_model.state_dict() , UpperCamelCase__ ) model.load_state_dict(UpperCamelCase__ ) _a : List[str] = """http://images.cocodataset.org/val2017/000000039769.jpg""" _a : Optional[Any] = AutoImageProcessor.from_pretrained("""microsoft/{}""".format(swinva_name.replace("""_""" , """-""" ) ) ) _a : str = Image.open(requests.get(UpperCamelCase__ , stream=UpperCamelCase__ ).raw ) _a : Any = image_processor(images=UpperCamelCase__ , return_tensors="""pt""" ) _a : Optional[Any] = timm_model(inputs["""pixel_values"""] ) _a : Dict = model(**UpperCamelCase__ ).logits assert torch.allclose(UpperCamelCase__ , UpperCamelCase__ , atol=1e-3 ) print(F"""Saving model {swinva_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(UpperCamelCase__ ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(UpperCamelCase__ ) model.push_to_hub( repo_path_or_name=Path(UpperCamelCase__ , UpperCamelCase__ ) , organization="""nandwalritik""" , commit_message="""Add model""" , ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '--swinv2_name', default='swinv2_tiny_patch4_window8_256', type=str, help='Name of the Swinv2 timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) _snake_case = parser.parse_args() convert_swinva_checkpoint(args.swinva_name, args.pytorch_dump_folder_path)
294
"""simple docstring""" import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device _snake_case = False class UpperCamelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : Optional[Any] ) -> Optional[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self : Tuple ) -> List[Any]: _a : Tuple = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : List[str] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) _a : Optional[Any] = torch.manual_seed(0 ) _a : Union[str, Any] = pipe.dual_guided( prompt="""first prompt""" , image=UpperCAmelCase__ , text_to_image_strength=0.7_5 , generator=UpperCAmelCase__ , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(UpperCAmelCase__ ) _a : Dict = VersatileDiffusionPipeline.from_pretrained(UpperCAmelCase__ , torch_dtype=torch.floataa ) pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Optional[Any] = generator.manual_seed(0 ) _a : str = pipe.dual_guided( prompt="""first prompt""" , image=UpperCAmelCase__ , text_to_image_strength=0.7_5 , generator=UpperCAmelCase__ , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def _lowercase ( self : Optional[int] ) -> Optional[int]: _a : Optional[int] = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : int = """cyberpunk 2077""" _a : Tuple = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) _a : Tuple = torch.manual_seed(0 ) _a : Any = pipe.dual_guided( prompt=UpperCAmelCase__ , image=UpperCAmelCase__ , text_to_image_strength=0.7_5 , generator=UpperCAmelCase__ , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" , ).images _a : List[str] = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _a : Optional[int] = np.array([0.1_4_4_8, 0.1_6_1_9, 0.1_7_4_1, 0.1_0_8_6, 0.1_1_4_7, 0.1_1_2_8, 0.1_1_9_9, 0.1_1_6_5, 0.1_0_0_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 _a : int = """A painting of a squirrel eating a burger """ _a : Tuple = torch.manual_seed(0 ) _a : Union[str, Any] = pipe.text_to_image( prompt=UpperCAmelCase__ , generator=UpperCAmelCase__ , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" ).images _a : int = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _a : int = np.array([0.3_3_6_7, 0.3_1_6_9, 0.2_6_5_6, 0.3_8_7_0, 0.4_7_9_0, 0.3_7_9_6, 0.4_0_0_9, 0.4_8_7_8, 0.4_7_7_8] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 _a : str = pipe.image_variation(UpperCAmelCase__ , generator=UpperCAmelCase__ , output_type="""numpy""" ).images _a : str = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _a : Optional[Any] = np.array([0.3_0_7_6, 0.3_1_2_3, 0.3_2_8_4, 0.3_7_8_2, 0.3_7_7_0, 0.3_8_9_4, 0.4_2_9_7, 0.4_3_3_1, 0.4_4_5_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1
294
1
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyInpaintPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.text_encoder import MCLIPConfig, MultilingualCLIP from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class UpperCamelCase ( snake_case_ , unittest.TestCase ): UpperCamelCase : Tuple = KandinskyInpaintPipeline UpperCamelCase : str = ['''prompt''', '''image_embeds''', '''negative_image_embeds''', '''image''', '''mask_image'''] UpperCamelCase : Tuple = [ '''prompt''', '''negative_prompt''', '''image_embeds''', '''negative_image_embeds''', '''image''', '''mask_image''', ] UpperCamelCase : Tuple = [ '''generator''', '''height''', '''width''', '''latents''', '''guidance_scale''', '''negative_prompt''', '''num_inference_steps''', '''return_dict''', '''guidance_scale''', '''num_images_per_prompt''', '''output_type''', '''return_dict''', ] UpperCamelCase : Any = False @property def _lowercase ( self : Union[str, Any] ) -> Any: return 32 @property def _lowercase ( self : str ) -> int: return 32 @property def _lowercase ( self : int ) -> Tuple: return self.time_input_dim @property def _lowercase ( self : Dict ) -> Union[str, Any]: return self.time_input_dim * 4 @property def _lowercase ( self : Optional[int] ) -> Any: return 100 @property def _lowercase ( self : Optional[int] ) -> Dict: _a : int = XLMRobertaTokenizerFast.from_pretrained("""YiYiXu/tiny-random-mclip-base""" ) return tokenizer @property def _lowercase ( self : Any ) -> Any: torch.manual_seed(0 ) _a : List[Any] = MCLIPConfig( numDims=self.cross_attention_dim , transformerDimensions=self.text_embedder_hidden_size , hidden_size=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=1005 , ) _a : Tuple = MultilingualCLIP(UpperCAmelCase__ ) _a : Union[str, Any] = text_encoder.eval() return text_encoder @property def _lowercase ( self : List[str] ) -> Tuple: torch.manual_seed(0 ) _a : Optional[int] = { """in_channels""": 9, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """text_image""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """text_image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } _a : str = UNetaDConditionModel(**UpperCAmelCase__ ) return model @property def _lowercase ( self : List[str] ) -> List[str]: return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def _lowercase ( self : Union[str, Any] ) -> Any: torch.manual_seed(0 ) _a : List[str] = VQModel(**self.dummy_movq_kwargs ) return model def _lowercase ( self : Union[str, Any] ) -> Tuple: _a : List[Any] = self.dummy_text_encoder _a : Dict = self.dummy_tokenizer _a : Union[str, Any] = self.dummy_unet _a : List[Any] = self.dummy_movq _a : str = DDIMScheduler( num_train_timesteps=1000 , beta_schedule="""linear""" , beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , clip_sample=UpperCAmelCase__ , set_alpha_to_one=UpperCAmelCase__ , steps_offset=1 , prediction_type="""epsilon""" , thresholding=UpperCAmelCase__ , ) _a : Union[str, Any] = { """text_encoder""": text_encoder, """tokenizer""": tokenizer, """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def _lowercase ( self : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : Any=0 ) -> Tuple: _a : Optional[int] = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(UpperCAmelCase__ ) ).to(UpperCAmelCase__ ) _a : Optional[Any] = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(seed + 1 ) ).to(UpperCAmelCase__ ) # create init_image _a : int = floats_tensor((1, 3, 64, 64) , rng=random.Random(UpperCAmelCase__ ) ).to(UpperCAmelCase__ ) _a : Dict = image.cpu().permute(0 , 2 , 3 , 1 )[0] _a : str = Image.fromarray(np.uinta(UpperCAmelCase__ ) ).convert("""RGB""" ).resize((256, 256) ) # create mask _a : Optional[Any] = np.ones((64, 64) , dtype=np.floataa ) _a : Union[str, Any] = 0 if str(UpperCAmelCase__ ).startswith("""mps""" ): _a : Optional[Any] = torch.manual_seed(UpperCAmelCase__ ) else: _a : int = torch.Generator(device=UpperCAmelCase__ ).manual_seed(UpperCAmelCase__ ) _a : Any = { """prompt""": """horse""", """image""": init_image, """mask_image""": mask, """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """generator""": generator, """height""": 64, """width""": 64, """num_inference_steps""": 2, """guidance_scale""": 4.0, """output_type""": """np""", } return inputs def _lowercase ( self : Any ) -> Union[str, Any]: _a : Tuple = """cpu""" _a : int = self.get_dummy_components() _a : Optional[int] = self.pipeline_class(**UpperCAmelCase__ ) _a : Optional[int] = pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : str = pipe(**self.get_dummy_inputs(UpperCAmelCase__ ) ) _a : Dict = output.images _a : Optional[int] = pipe( **self.get_dummy_inputs(UpperCAmelCase__ ) , return_dict=UpperCAmelCase__ , )[0] _a : List[Any] = image[0, -3:, -3:, -1] _a : Tuple = image_from_tuple[0, -3:, -3:, -1] print(f"""image.shape {image.shape}""" ) assert image.shape == (1, 64, 64, 3) _a : Optional[int] = np.array( [0.8_3_2_6_9_1_9, 0.7_3_7_9_0_4_6_7, 0.2_0_9_1_8_5_8_1, 0.9_3_0_9_6_1_2, 0.5_5_1_1_7_9_1, 0.4_3_7_1_3_3_2_8, 0.5_5_1_3_3_2_1, 0.4_9_9_2_2_9_3_4, 0.5_9_4_9_7_7_8_6] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), f""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), f""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" def _lowercase ( self : Tuple ) -> Optional[int]: super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : Tuple ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self : Optional[Any] ) -> str: _a : Optional[int] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinsky/kandinsky_inpaint_cat_with_hat_fp16.npy""" ) _a : int = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinsky/cat.png""" ) _a : int = np.ones((768, 768) , dtype=np.floataa ) _a : List[Any] = 0 _a : Optional[Any] = """a hat""" _a : int = KandinskyPriorPipeline.from_pretrained( """kandinsky-community/kandinsky-2-1-prior""" , torch_dtype=torch.floataa ) pipe_prior.to(UpperCAmelCase__ ) _a : int = KandinskyInpaintPipeline.from_pretrained( """kandinsky-community/kandinsky-2-1-inpaint""" , torch_dtype=torch.floataa ) _a : Tuple = pipeline.to(UpperCAmelCase__ ) pipeline.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Optional[Any] = torch.Generator(device="""cpu""" ).manual_seed(0 ) _a , _a : int = pipe_prior( UpperCAmelCase__ , generator=UpperCAmelCase__ , num_inference_steps=5 , negative_prompt="""""" , ).to_tuple() _a : Dict = pipeline( UpperCAmelCase__ , image=UpperCAmelCase__ , mask_image=UpperCAmelCase__ , image_embeds=UpperCAmelCase__ , negative_image_embeds=UpperCAmelCase__ , generator=UpperCAmelCase__ , num_inference_steps=100 , height=768 , width=768 , output_type="""np""" , ) _a : str = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(UpperCAmelCase__ , UpperCAmelCase__ )
294
"""simple docstring""" import json import os from typing import Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = {'vocab_file': 'vocab.json'} _snake_case = { 'vocab_file': { 'mgp-str': 'https://huggingface.co/alibaba-damo/mgp-str-base/blob/main/vocab.json', } } _snake_case = {'mgp-str': 27} class UpperCamelCase ( snake_case_ ): UpperCamelCase : List[str] = VOCAB_FILES_NAMES UpperCamelCase : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : Optional[int] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : List[Any]="[GO]" , UpperCAmelCase__ : Tuple="[GO]" , UpperCAmelCase__ : Optional[int]="[s]" , UpperCAmelCase__ : int="[GO]" , **UpperCAmelCase__ : Dict ) -> int: super().__init__( unk_token=UpperCAmelCase__ , bos_token=UpperCAmelCase__ , eos_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , **UpperCAmelCase__ , ) with open(UpperCAmelCase__ , encoding="""utf-8""" ) as vocab_handle: _a : int = json.load(UpperCAmelCase__ ) _a : Optional[int] = {v: k for k, v in self.vocab.items()} @property def _lowercase ( self : Dict ) -> Union[str, Any]: return len(self.vocab ) def _lowercase ( self : Union[str, Any] ) -> str: return dict(self.vocab , **self.added_tokens_encoder ) def _lowercase ( self : Dict , UpperCAmelCase__ : str ) -> Union[str, Any]: _a : Tuple = [] for s in text: char_tokens.extend(UpperCAmelCase__ ) return char_tokens def _lowercase ( self : List[Any] , UpperCAmelCase__ : str ) -> Dict: return self.vocab.get(UpperCAmelCase__ , self.vocab.get(self.unk_token ) ) def _lowercase ( self : Optional[int] , UpperCAmelCase__ : Tuple ) -> List[Any]: return self.decoder.get(UpperCAmelCase__ ) def _lowercase ( self : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(UpperCAmelCase__ ): logger.error("""Vocabulary path ({}) should be a directory""".format(UpperCAmelCase__ ) ) return _a : Tuple = os.path.join( UpperCAmelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) with open(UpperCAmelCase__ , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.vocab , indent=2 , sort_keys=UpperCAmelCase__ , ensure_ascii=UpperCAmelCase__ ) + """\n""" ) return (vocab_file,)
294
1
"""simple docstring""" def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , ): '''simple docstring''' _a : Optional[Any] = [redshift, radiation_density, matter_density, dark_energy] if any(p < 0 for p in parameters ): raise ValueError("""All input parameters must be positive""" ) if any(p > 1 for p in parameters[1:4] ): raise ValueError("""Relative densities cannot be greater than one""" ) else: _a : Tuple = 1 - (matter_density + radiation_density + dark_energy) _a : int = ( radiation_density * (redshift + 1) ** 4 + matter_density * (redshift + 1) ** 3 + curvature * (redshift + 1) ** 2 + dark_energy ) _a : List[str] = hubble_constant * e_a ** (1 / 2) return hubble if __name__ == "__main__": import doctest # run doctest doctest.testmod() # demo LCDM approximation _snake_case = 0.3 print( hubble_parameter( hubble_constant=68.3, radiation_density=1e-4, matter_density=matter_density, dark_energy=1 - matter_density, redshift=0, ) )
294
"""simple docstring""" import tempfile import torch from diffusers import IPNDMScheduler from .test_schedulers import SchedulerCommonTest class UpperCamelCase ( snake_case_ ): UpperCamelCase : int = (IPNDMScheduler,) UpperCamelCase : int = (('''num_inference_steps''', 50),) def _lowercase ( self : Union[str, Any] , **UpperCAmelCase__ : Tuple ) -> int: _a : Optional[int] = {"""num_train_timesteps""": 1000} config.update(**UpperCAmelCase__ ) return config def _lowercase ( self : Dict , UpperCAmelCase__ : Any=0 , **UpperCAmelCase__ : Optional[Any] ) -> Union[str, Any]: _a : Optional[int] = dict(self.forward_default_kwargs ) _a : Dict = kwargs.pop("""num_inference_steps""" , UpperCAmelCase__ ) _a : Optional[Any] = self.dummy_sample _a : Union[str, Any] = 0.1 * sample _a : Union[str, Any] = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] for scheduler_class in self.scheduler_classes: _a : Optional[int] = self.get_scheduler_config(**UpperCAmelCase__ ) _a : Union[str, Any] = scheduler_class(**UpperCAmelCase__ ) scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residuals _a : Any = dummy_past_residuals[:] if time_step is None: _a : str = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(UpperCAmelCase__ ) _a : Union[str, Any] = scheduler_class.from_pretrained(UpperCAmelCase__ ) new_scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residuals _a : Optional[Any] = dummy_past_residuals[:] _a : List[Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : str = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" _a : Optional[int] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : Tuple = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def _lowercase ( self : Tuple ) -> List[str]: pass def _lowercase ( self : Optional[int] , UpperCAmelCase__ : List[str]=0 , **UpperCAmelCase__ : Optional[Any] ) -> List[Any]: _a : Optional[Any] = dict(self.forward_default_kwargs ) _a : Optional[Any] = kwargs.pop("""num_inference_steps""" , UpperCAmelCase__ ) _a : Optional[Any] = self.dummy_sample _a : List[Any] = 0.1 * sample _a : Optional[Any] = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] for scheduler_class in self.scheduler_classes: _a : Union[str, Any] = self.get_scheduler_config() _a : Optional[Any] = scheduler_class(**UpperCAmelCase__ ) scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residuals (must be after setting timesteps) _a : Any = dummy_past_residuals[:] if time_step is None: _a : List[Any] = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(UpperCAmelCase__ ) _a : Any = scheduler_class.from_pretrained(UpperCAmelCase__ ) # copy over dummy past residuals new_scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residual (must be after setting timesteps) _a : Optional[Any] = dummy_past_residuals[:] _a : List[str] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : Tuple = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" _a : Union[str, Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : int = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def _lowercase ( self : str , **UpperCAmelCase__ : Any ) -> List[str]: _a : Optional[int] = self.scheduler_classes[0] _a : Optional[Any] = self.get_scheduler_config(**UpperCAmelCase__ ) _a : Union[str, Any] = scheduler_class(**UpperCAmelCase__ ) _a : int = 10 _a : List[Any] = self.dummy_model() _a : str = self.dummy_sample_deter scheduler.set_timesteps(UpperCAmelCase__ ) for i, t in enumerate(scheduler.timesteps ): _a : str = model(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ).prev_sample for i, t in enumerate(scheduler.timesteps ): _a : Union[str, Any] = model(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Any = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ).prev_sample return sample def _lowercase ( self : int ) -> str: _a : Dict = dict(self.forward_default_kwargs ) _a : int = kwargs.pop("""num_inference_steps""" , UpperCAmelCase__ ) for scheduler_class in self.scheduler_classes: _a : Optional[int] = self.get_scheduler_config() _a : Tuple = scheduler_class(**UpperCAmelCase__ ) _a : Tuple = self.dummy_sample _a : Optional[Any] = 0.1 * sample if num_inference_steps is not None and hasattr(UpperCAmelCase__ , """set_timesteps""" ): scheduler.set_timesteps(UpperCAmelCase__ ) elif num_inference_steps is not None and not hasattr(UpperCAmelCase__ , """set_timesteps""" ): _a : List[str] = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) _a : Union[str, Any] = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] _a : Optional[Any] = dummy_past_residuals[:] _a : Optional[Any] = scheduler.timesteps[5] _a : str = scheduler.timesteps[6] _a : Optional[int] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : Union[str, Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) _a : Tuple = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : List[str] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def _lowercase ( self : List[str] ) -> List[str]: for timesteps in [100, 1000]: self.check_over_configs(num_train_timesteps=UpperCAmelCase__ , time_step=UpperCAmelCase__ ) def _lowercase ( self : List[str] ) -> List[str]: for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 100] ): self.check_over_forward(num_inference_steps=UpperCAmelCase__ , time_step=UpperCAmelCase__ ) def _lowercase ( self : int ) -> List[Any]: _a : str = self.full_loop() _a : List[Any] = torch.mean(torch.abs(UpperCAmelCase__ ) ) assert abs(result_mean.item() - 2540529 ) < 10
294
1
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class UpperCamelCase ( unittest.TestCase ): def __init__( self : Optional[int] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Optional[Any]=7 , UpperCAmelCase__ : List[str]=3 , UpperCAmelCase__ : Optional[int]=18 , UpperCAmelCase__ : Dict=30 , UpperCAmelCase__ : int=400 , UpperCAmelCase__ : str=True , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Optional[int]=True , UpperCAmelCase__ : Tuple=None , UpperCAmelCase__ : str=True , ) -> Dict: _a : Union[str, Any] = size if size is not None else {"""shortest_edge""": 20} _a : Optional[int] = crop_size if crop_size is not None else {"""height""": 18, """width""": 18} _a : Optional[int] = parent _a : str = batch_size _a : Optional[int] = num_channels _a : str = image_size _a : List[Any] = min_resolution _a : Any = max_resolution _a : List[str] = do_resize _a : str = size _a : Optional[Any] = do_center_crop _a : str = crop_size _a : List[Any] = do_flip_channel_order def _lowercase ( self : List[Any] ) -> List[str]: return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_flip_channel_order": self.do_flip_channel_order, } @require_torch @require_vision class UpperCamelCase ( snake_case_ , unittest.TestCase ): UpperCamelCase : Dict = MobileViTImageProcessor if is_vision_available() else None def _lowercase ( self : Any ) -> Union[str, Any]: _a : Union[str, Any] = MobileViTImageProcessingTester(self ) @property def _lowercase ( self : Dict ) -> str: return self.image_processor_tester.prepare_image_processor_dict() def _lowercase ( self : str ) -> Tuple: _a : Any = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCAmelCase__ , """do_resize""" ) ) self.assertTrue(hasattr(UpperCAmelCase__ , """size""" ) ) self.assertTrue(hasattr(UpperCAmelCase__ , """do_center_crop""" ) ) self.assertTrue(hasattr(UpperCAmelCase__ , """center_crop""" ) ) self.assertTrue(hasattr(UpperCAmelCase__ , """do_flip_channel_order""" ) ) def _lowercase ( self : str ) -> str: _a : int = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""shortest_edge""": 20} ) self.assertEqual(image_processor.crop_size , {"""height""": 18, """width""": 18} ) _a : List[str] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"""shortest_edge""": 42} ) self.assertEqual(image_processor.crop_size , {"""height""": 84, """width""": 84} ) def _lowercase ( self : Optional[int] ) -> List[Any]: pass def _lowercase ( self : Optional[Any] ) -> Tuple: # Initialize image_processing _a : str = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _a : Tuple = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , Image.Image ) # Test not batched input _a : List[Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched _a : str = image_processing(UpperCAmelCase__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def _lowercase ( self : int ) -> Tuple: # Initialize image_processing _a : Any = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _a : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ , numpify=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , np.ndarray ) # Test not batched input _a : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched _a : Any = image_processing(UpperCAmelCase__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def _lowercase ( self : List[str] ) -> Dict: # Initialize image_processing _a : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _a : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ , torchify=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , torch.Tensor ) # Test not batched input _a : int = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched _a : str = image_processing(UpperCAmelCase__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , )
294
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging _snake_case = logging.get_logger(__name__) _snake_case = {'vocab_file': 'spiece.model'} _snake_case = { 'vocab_file': { 'TsinghuaAI/CPM-Generate': 'https://huggingface.co/TsinghuaAI/CPM-Generate/resolve/main/spiece.model', } } class UpperCamelCase ( snake_case_ ): def __init__( self : Tuple , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : int=False , UpperCAmelCase__ : Dict=True , UpperCAmelCase__ : Union[str, Any]=False , UpperCAmelCase__ : Dict="<s>" , UpperCAmelCase__ : Any="</s>" , UpperCAmelCase__ : Any="<unk>" , UpperCAmelCase__ : int="<sep>" , UpperCAmelCase__ : Tuple="<pad>" , UpperCAmelCase__ : Any="<cls>" , UpperCAmelCase__ : Optional[Any]="<mask>" , UpperCAmelCase__ : int=["<eop>", "<eod>"] , UpperCAmelCase__ : Optional[Dict[str, Any]] = None , **UpperCAmelCase__ : List[str] , ) -> None: _a : Optional[int] = AddedToken(UpperCAmelCase__ , lstrip=UpperCAmelCase__ , rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else mask_token _a : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=UpperCAmelCase__ , remove_space=UpperCAmelCase__ , keep_accents=UpperCAmelCase__ , bos_token=UpperCAmelCase__ , eos_token=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , sep_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , cls_token=UpperCAmelCase__ , mask_token=UpperCAmelCase__ , additional_special_tokens=UpperCAmelCase__ , sp_model_kwargs=self.sp_model_kwargs , **UpperCAmelCase__ , ) _a : Optional[Any] = 3 _a : Tuple = do_lower_case _a : Tuple = remove_space _a : Tuple = keep_accents _a : Tuple = vocab_file _a : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCAmelCase__ ) try: import jieba except ModuleNotFoundError as error: raise error.__class__( """You need to install jieba to use CpmTokenizer or CpmTokenizerFast. """ """See https://pypi.org/project/jieba/ for installation.""" ) _a : int = jieba _a : Tuple = str.maketrans(""" \n""" , """\u2582\u2583""" ) @property # Copied from transformers.models.xlnet.tokenization_xlnet.XLNetTokenizer.vocab_size def _lowercase ( self : Optional[Any] ) -> Any: return len(self.sp_model ) def _lowercase ( self : str ) -> Union[str, Any]: _a : int = {self.convert_ids_to_tokens(UpperCAmelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Tuple ) -> List[str]: _a : Tuple = self.__dict__.copy() _a : Tuple = None return state def __setstate__( self : Any , UpperCAmelCase__ : Dict ) -> Dict: _a : Tuple = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): _a : Tuple = {} _a : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _lowercase ( self : List[str] , UpperCAmelCase__ : Union[str, Any] ) -> Dict: if self.remove_space: _a : Optional[int] = """ """.join(inputs.strip().split() ) else: _a : List[Any] = inputs _a : int = outputs.replace("""``""" , """\"""" ).replace("""''""" , """\"""" ) if not self.keep_accents: _a : Optional[Any] = unicodedata.normalize("""NFKD""" , UpperCAmelCase__ ) _a : Dict = """""".join([c for c in outputs if not unicodedata.combining(UpperCAmelCase__ )] ) if self.do_lower_case: _a : Union[str, Any] = outputs.lower() return outputs def _lowercase ( self : List[str] , UpperCAmelCase__ : str ) -> List[str]: _a : str = self.preprocess_text(UpperCAmelCase__ ) _a : Dict = self.sp_model.encode(UpperCAmelCase__ , out_type=UpperCAmelCase__ ) _a : Union[str, Any] = [] for piece in pieces: if len(UpperCAmelCase__ ) > 1 and piece[-1] == str(""",""" ) and piece[-2].isdigit(): _a : Dict = self.sp_model.EncodeAsPieces(piece[:-1].replace(UpperCAmelCase__ , """""" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: _a : Dict = cur_pieces[1:] else: _a : Any = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(UpperCAmelCase__ ) else: new_pieces.append(UpperCAmelCase__ ) return new_pieces def _lowercase ( self : Union[str, Any] , UpperCAmelCase__ : int ) -> int: return self.sp_model.PieceToId(UpperCAmelCase__ ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : Optional[Any] ) -> Any: return self.sp_model.IdToPiece(UpperCAmelCase__ ) def _lowercase ( self : Any , UpperCAmelCase__ : Any ) -> Dict: _a : Dict = """""".join(UpperCAmelCase__ ).replace(UpperCAmelCase__ , """ """ ).strip() return out_string def _lowercase ( self : Any , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ) -> List[int]: _a : Optional[Any] = [self.sep_token_id] _a : Dict = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _lowercase ( self : Tuple , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None , UpperCAmelCase__ : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase__ , token_ids_a=UpperCAmelCase__ , already_has_special_tokens=UpperCAmelCase__ ) if token_ids_a is not None: return ([0] * len(UpperCAmelCase__ )) + [1] + ([0] * len(UpperCAmelCase__ )) + [1, 1] return ([0] * len(UpperCAmelCase__ )) + [1, 1] def _lowercase ( self : str , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ) -> List[int]: _a : Any = [self.sep_token_id] _a : Optional[Any] = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _lowercase ( self : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(UpperCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return _a : Union[str, Any] = os.path.join( UpperCAmelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCAmelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCAmelCase__ , """wb""" ) as fi: _a : Optional[Any] = self.sp_model.serialized_model_proto() fi.write(UpperCAmelCase__ ) return (out_vocab_file,) def _lowercase ( self : Any , *UpperCAmelCase__ : List[str] , **UpperCAmelCase__ : List[str] ) -> List[str]: _a : Tuple = super()._decode(*UpperCAmelCase__ , **UpperCAmelCase__ ) _a : Optional[Any] = text.replace(""" """ , """""" ).replace("""\u2582""" , """ """ ).replace("""\u2583""" , """\n""" ) return text
294
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_videomae import VideoMAEImageProcessor _snake_case = logging.get_logger(__name__) class UpperCamelCase ( snake_case_ ): def __init__( self : List[Any] , *UpperCAmelCase__ : Tuple , **UpperCAmelCase__ : Tuple ) -> None: warnings.warn( """The class VideoMAEFeatureExtractor is deprecated and will be removed in version 5 of Transformers.""" """ Please use VideoMAEImageProcessor instead.""" , UpperCAmelCase__ , ) super().__init__(*UpperCAmelCase__ , **UpperCAmelCase__ )
294
"""simple docstring""" import inspect import unittest import torch import torch.nn as nn from accelerate.hooks import ( AlignDevicesHook, ModelHook, SequentialHook, add_hook_to_module, attach_align_device_hook, remove_hook_from_module, remove_hook_from_submodules, ) from accelerate.test_utils import require_multi_gpu class UpperCamelCase ( nn.Module ): def __init__( self : Union[str, Any] ) -> int: super().__init__() _a : Optional[Any] = nn.Linear(3 , 4 ) _a : Tuple = nn.BatchNormad(4 ) _a : Dict = nn.Linear(4 , 5 ) def _lowercase ( self : Optional[int] , UpperCAmelCase__ : List[str] ) -> int: return self.lineara(self.batchnorm(self.lineara(UpperCAmelCase__ ) ) ) class UpperCamelCase ( snake_case_ ): def _lowercase ( self : Any , UpperCAmelCase__ : Any , *UpperCAmelCase__ : List[str] , **UpperCAmelCase__ : Optional[int] ) -> Optional[int]: return (args[0] + 1,) + args[1:], kwargs class UpperCamelCase ( snake_case_ ): def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Optional[Any] ) -> List[str]: return output + 1 class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : Dict ) -> str: _a : List[Any] = ModelForTest() _a : str = ModelHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) self.assertEqual(test_model._hf_hook , UpperCAmelCase__ ) self.assertTrue(hasattr(UpperCAmelCase__ , """_old_forward""" ) ) # Check adding the hook did not change the name or the signature self.assertEqual(test_model.forward.__name__ , """forward""" ) self.assertListEqual(list(inspect.signature(test_model.forward ).parameters ) , ["""x"""] ) remove_hook_from_module(UpperCAmelCase__ ) self.assertFalse(hasattr(UpperCAmelCase__ , """_hf_hook""" ) ) self.assertFalse(hasattr(UpperCAmelCase__ , """_old_forward""" ) ) def _lowercase ( self : Optional[int] ) -> Optional[int]: _a : Dict = ModelForTest() _a : Dict = ModelHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ , append=UpperCAmelCase__ ) self.assertEqual(isinstance(test_model._hf_hook , UpperCAmelCase__ ) , UpperCAmelCase__ ) self.assertEqual(len(test_model._hf_hook.hooks ) , 2 ) self.assertTrue(hasattr(UpperCAmelCase__ , """_old_forward""" ) ) # Check adding the hook did not change the name or the signature self.assertEqual(test_model.forward.__name__ , """forward""" ) self.assertListEqual(list(inspect.signature(test_model.forward ).parameters ) , ["""x"""] ) remove_hook_from_module(UpperCAmelCase__ ) self.assertFalse(hasattr(UpperCAmelCase__ , """_hf_hook""" ) ) self.assertFalse(hasattr(UpperCAmelCase__ , """_old_forward""" ) ) def _lowercase ( self : Dict ) -> int: _a : str = ModelForTest() _a : List[Any] = torch.randn(2 , 3 ) _a : Optional[Any] = test_model(x + 1 ) _a : str = test_model(x + 2 ) _a : Union[str, Any] = PreForwardHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Tuple = test_model(UpperCAmelCase__ ) self.assertTrue(torch.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1E-5 ) ) # Attaching a hook to a model when it already has one replaces, does not chain _a : int = PreForwardHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : str = test_model(UpperCAmelCase__ ) self.assertTrue(torch.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1E-5 ) ) # You need to use the sequential hook to chain two or more hooks _a : int = SequentialHook(PreForwardHook() , PreForwardHook() ) add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Tuple = test_model(UpperCAmelCase__ ) assert torch.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1E-5 ) def _lowercase ( self : Tuple ) -> int: _a : Tuple = ModelForTest() _a : Union[str, Any] = torch.randn(2 , 3 ) _a : Optional[int] = test_model(UpperCAmelCase__ ) _a : int = PostForwardHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[str] = test_model(UpperCAmelCase__ ) self.assertTrue(torch.allclose(UpperCAmelCase__ , output + 1 , atol=1E-5 ) ) # Attaching a hook to a model when it already has one replaces, does not chain _a : List[Any] = PostForwardHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Dict = test_model(UpperCAmelCase__ ) self.assertTrue(torch.allclose(UpperCAmelCase__ , output + 1 , atol=1E-5 ) ) # You need to use the sequential hook to chain two or more hooks _a : Any = SequentialHook(PostForwardHook() , PostForwardHook() ) add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Optional[int] = test_model(UpperCAmelCase__ ) assert torch.allclose(UpperCAmelCase__ , output + 2 , atol=1E-5 ) def _lowercase ( self : Dict ) -> Optional[Any]: _a : Any = ModelForTest() _a : List[Any] = torch.randn(2 , 3 ) _a : Dict = test_model(UpperCAmelCase__ ) _a : Any = PostForwardHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[str] = test_model(UpperCAmelCase__ ) self.assertTrue(torch.allclose(UpperCAmelCase__ , output + 1 ) ) self.assertTrue(outputa.requires_grad ) _a : Any = True _a : Union[str, Any] = test_model(UpperCAmelCase__ ) self.assertFalse(outputa.requires_grad ) @require_multi_gpu def _lowercase ( self : Optional[Any] ) -> str: _a : List[Any] = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices add_hook_to_module(model.lineara , AlignDevicesHook(execution_device=0 ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(execution_device=0 ) ) add_hook_to_module(model.lineara , AlignDevicesHook(execution_device=1 ) ) self.assertEqual(model.lineara.weight.device , torch.device(0 ) ) self.assertEqual(model.batchnorm.weight.device , torch.device(0 ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device(0 ) ) self.assertEqual(model.lineara.weight.device , torch.device(1 ) ) # We can still make a forward pass. The input does not need to be on any particular device _a : Optional[int] = torch.randn(2 , 3 ) _a : Any = model(UpperCAmelCase__ ) self.assertEqual(output.device , torch.device(1 ) ) # We can add a general hook to put back output on same device as input. add_hook_to_module(UpperCAmelCase__ , AlignDevicesHook(io_same_device=UpperCAmelCase__ ) ) _a : str = torch.randn(2 , 3 ).to(0 ) _a : Union[str, Any] = model(UpperCAmelCase__ ) self.assertEqual(output.device , torch.device(0 ) ) def _lowercase ( self : str ) -> Union[str, Any]: _a : int = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices _a : List[Any] = {"""execution_device""": 0 if torch.cuda.is_available() else """cpu""", """offload""": True} add_hook_to_module(model.lineara , AlignDevicesHook(**UpperCAmelCase__ ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(**UpperCAmelCase__ ) ) add_hook_to_module(model.lineara , AlignDevicesHook(**UpperCAmelCase__ ) ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) # Buffers are not included in the offload by default, so are on the execution device _a : Dict = torch.device(hook_kwargs["""execution_device"""] ) self.assertEqual(model.batchnorm.running_mean.device , UpperCAmelCase__ ) _a : int = torch.randn(2 , 3 ) _a : str = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_module(model.lineara ) remove_hook_from_module(model.batchnorm ) remove_hook_from_module(model.lineara ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # Now test with buffers included in the offload _a : List[str] = { """execution_device""": 0 if torch.cuda.is_available() else """cpu""", """offload""": True, """offload_buffers""": True, } add_hook_to_module(model.lineara , AlignDevicesHook(**UpperCAmelCase__ ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(**UpperCAmelCase__ ) ) add_hook_to_module(model.lineara , AlignDevicesHook(**UpperCAmelCase__ ) ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device("""meta""" ) ) _a : Tuple = torch.randn(2 , 3 ) _a : Union[str, Any] = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_module(model.lineara ) remove_hook_from_module(model.batchnorm ) remove_hook_from_module(model.lineara ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) def _lowercase ( self : Tuple ) -> List[str]: _a : str = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices _a : Union[str, Any] = 0 if torch.cuda.is_available() else """cpu""" attach_align_device_hook(UpperCAmelCase__ , execution_device=UpperCAmelCase__ , offload=UpperCAmelCase__ ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) # Buffers are not included in the offload by default, so are on the execution device _a : Dict = torch.device(UpperCAmelCase__ ) self.assertEqual(model.batchnorm.running_mean.device , UpperCAmelCase__ ) _a : Union[str, Any] = torch.randn(2 , 3 ) _a : List[Any] = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(UpperCAmelCase__ ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # Now test with buffers included in the offload attach_align_device_hook(UpperCAmelCase__ , execution_device=UpperCAmelCase__ , offload=UpperCAmelCase__ , offload_buffers=UpperCAmelCase__ ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device("""meta""" ) ) _a : List[str] = torch.randn(2 , 3 ) _a : Union[str, Any] = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(UpperCAmelCase__ ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) def _lowercase ( self : Dict ) -> str: _a : Optional[Any] = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices _a : str = 0 if torch.cuda.is_available() else """cpu""" attach_align_device_hook( UpperCAmelCase__ , execution_device=UpperCAmelCase__ , offload=UpperCAmelCase__ , weights_map=model.state_dict() ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) # Buffers are not included in the offload by default, so are on the execution device _a : Union[str, Any] = torch.device(UpperCAmelCase__ ) self.assertEqual(model.batchnorm.running_mean.device , UpperCAmelCase__ ) _a : Union[str, Any] = torch.randn(2 , 3 ) _a : int = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(UpperCAmelCase__ ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # Now test with buffers included in the offload attach_align_device_hook( UpperCAmelCase__ , execution_device=UpperCAmelCase__ , offload=UpperCAmelCase__ , weights_map=model.state_dict() , offload_buffers=UpperCAmelCase__ , ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device("""meta""" ) ) _a : Any = torch.randn(2 , 3 ) _a : int = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(UpperCAmelCase__ ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) )
294
1
"""simple docstring""" import requests _snake_case = 'https://newsapi.org/v1/articles?source=bbc-news&sortBy=top&apiKey=' def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' # fetching a list of articles in json format _a : Dict = requests.get(_NEWS_API + bbc_news_api_key ).json() # each article in the list is a dict for i, article in enumerate(bbc_news_page["""articles"""] , 1 ): print(F"""{i}.) {article['title']}""" ) if __name__ == "__main__": fetch_bbc_news(bbc_news_api_key='<Your BBC News API key goes here>')
294
"""simple docstring""" from __future__ import annotations def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' print(F"""Vertex\tShortest Distance from vertex {src}""" ) for i, d in enumerate(UpperCamelCase__ ): print(F"""{i}\t\t{d}""" ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' for j in range(UpperCamelCase__ ): _a , _a , _a : List[str] = (graph[j][k] for k in ["""src""", """dst""", """weight"""]) if distance[u] != float("""inf""" ) and distance[u] + w < distance[v]: return True return False def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Dict = [float("""inf""" )] * vertex_count _a : Any = 0.0 for _ in range(vertex_count - 1 ): for j in range(UpperCamelCase__ ): _a , _a , _a : List[Any] = (graph[j][k] for k in ["""src""", """dst""", """weight"""]) if distance[u] != float("""inf""" ) and distance[u] + w < distance[v]: _a : Any = distance[u] + w _a : Union[str, Any] = check_negative_cycle(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) if negative_cycle_exists: raise Exception("""Negative cycle found""" ) return distance if __name__ == "__main__": import doctest doctest.testmod() _snake_case = int(input('Enter number of vertices: ').strip()) _snake_case = int(input('Enter number of edges: ').strip()) _snake_case = [{} for _ in range(E)] for i in range(E): print('Edge ', i + 1) _snake_case , _snake_case , _snake_case = ( int(x) for x in input('Enter source, destination, weight: ').strip().split(' ') ) _snake_case = {'src': src, 'dst': dest, 'weight': weight} _snake_case = int(input('\nEnter shortest path source:').strip()) _snake_case = bellman_ford(graph, V, E, source) print_distance(shortest_distance, 0)
294
1
"""simple docstring""" import os from itertools import chain from random import randrange, shuffle import pytest from .sola import PokerHand _snake_case = ( '4S 3H 2C 7S 5H', '9D 8H 2C 6S 7H', '2D 6D 9D TH 7D', 'TC 8C 2S JH 6C', 'JH 8S TH AH QH', 'TS KS 5S 9S AC', 'KD 6S 9D TH AD', 'KS 8D 4D 9S 4S', # pair '8C 4S KH JS 4D', # pair 'QH 8H KD JH 8S', # pair 'KC 4H KS 2H 8D', # pair 'KD 4S KC 3H 8S', # pair 'AH 8S AS KC JH', # pair '3H 4C 4H 3S 2H', # 2 pairs '5S 5D 2C KH KH', # 2 pairs '3C KH 5D 5S KH', # 2 pairs 'AS 3C KH AD KH', # 2 pairs '7C 7S 3S 7H 5S', # 3 of a kind '7C 7S KH 2H 7H', # 3 of a kind 'AC KH QH AH AS', # 3 of a kind '2H 4D 3C AS 5S', # straight (low ace) '3C 5C 4C 2C 6H', # straight '6S 8S 7S 5H 9H', # straight 'JS QS 9H TS KH', # straight 'QC KH TS JS AH', # straight (high ace) '8C 9C 5C 3C TC', # flush '3S 8S 9S 5S KS', # flush '4C 5C 9C 8C KC', # flush 'JH 8H AH KH QH', # flush '3D 2H 3H 2C 2D', # full house '2H 2C 3S 3H 3D', # full house 'KH KC 3S 3H 3D', # full house 'JC 6H JS JD JH', # 4 of a kind 'JC 7H JS JD JH', # 4 of a kind 'JC KH JS JD JH', # 4 of a kind '2S AS 4S 5S 3S', # straight flush (low ace) '2D 6D 3D 4D 5D', # straight flush '5C 6C 3C 7C 4C', # straight flush 'JH 9H TH KH QH', # straight flush 'JH AH TH KH QH', # royal flush (high ace straight flush) ) _snake_case = ( ('2H 3H 4H 5H 6H', 'KS AS TS QS JS', 'Loss'), ('2H 3H 4H 5H 6H', 'AS AD AC AH JD', 'Win'), ('AS AH 2H AD AC', 'JS JD JC JH 3D', 'Win'), ('2S AH 2H AS AC', 'JS JD JC JH AD', 'Loss'), ('2S AH 2H AS AC', '2H 3H 5H 6H 7H', 'Win'), ('AS 3S 4S 8S 2S', '2H 3H 5H 6H 7H', 'Win'), ('2H 3H 5H 6H 7H', '2S 3H 4H 5S 6C', 'Win'), ('2S 3H 4H 5S 6C', '3D 4C 5H 6H 2S', 'Tie'), ('2S 3H 4H 5S 6C', 'AH AC 5H 6H AS', 'Win'), ('2S 2H 4H 5S 4C', 'AH AC 5H 6H AS', 'Loss'), ('2S 2H 4H 5S 4C', 'AH AC 5H 6H 7S', 'Win'), ('6S AD 7H 4S AS', 'AH AC 5H 6H 7S', 'Loss'), ('2S AH 4H 5S KC', 'AH AC 5H 6H 7S', 'Loss'), ('2S 3H 6H 7S 9C', '7H 3C TH 6H 9S', 'Loss'), ('4S 5H 6H TS AC', '3S 5H 6H TS AC', 'Win'), ('2S AH 4H 5S 6C', 'AD 4C 5H 6H 2C', 'Tie'), ('AS AH 3H AD AC', 'AS AH 2H AD AC', 'Win'), ('AH AC 5H 5C QS', 'AH AC 5H 5C KS', 'Loss'), ('AH AC 5H 5C QS', 'KH KC 5H 5C QS', 'Win'), ('7C 7S KH 2H 7H', '3C 3S AH 2H 3H', 'Win'), ('3C 3S AH 2H 3H', '7C 7S KH 2H 7H', 'Loss'), ('6H 5H 4H 3H 2H', '5H 4H 3H 2H AH', 'Win'), ('5H 4H 3H 2H AH', '5H 4H 3H 2H AH', 'Tie'), ('5H 4H 3H 2H AH', '6H 5H 4H 3H 2H', 'Loss'), ('AH AD KS KC AC', 'AH KD KH AC KC', 'Win'), ('2H 4D 3C AS 5S', '2H 4D 3C 6S 5S', 'Loss'), ('2H 3S 3C 3H 2S', '3S 3C 2S 2H 2D', 'Win'), ('4D 6D 5D 2D JH', '3S 8S 3H TC KH', 'Loss'), ('4S 6C 8S 3S 7S', 'AD KS 2D 7D 7C', 'Loss'), ('6S 4C 7H 8C 3H', '5H JC AH 9D 9C', 'Loss'), ('9D 9H JH TC QH', '3C 2S JS 5C 7H', 'Win'), ('2H TC 8S AD 9S', '4H TS 7H 2C 5C', 'Win'), ('9D 3S 2C 7S 7C', 'JC TD 3C TC 9H', 'Loss'), ) _snake_case = ( ('2H 3H 4H 5H 6H', True), ('AS AH 2H AD AC', False), ('2H 3H 5H 6H 7H', True), ('KS AS TS QS JS', True), ('8H 9H QS JS TH', False), ('AS 3S 4S 8S 2S', True), ) _snake_case = ( ('2H 3H 4H 5H 6H', True), ('AS AH 2H AD AC', False), ('2H 3H 5H 6H 7H', False), ('KS AS TS QS JS', True), ('8H 9H QS JS TH', True), ) _snake_case = ( ('2H 4D 3C AS 5S', True, [5, 4, 3, 2, 14]), ('2H 5D 3C AS 5S', False, [14, 5, 5, 3, 2]), ('JH QD KC AS TS', False, [14, 13, 12, 11, 10]), ('9D 3S 2C 7S 7C', False, [9, 7, 7, 3, 2]), ) _snake_case = ( ('JH AH TH KH QH', 0), ('JH 9H TH KH QH', 0), ('JC KH JS JD JH', 7), ('KH KC 3S 3H 3D', 6), ('8C 9C 5C 3C TC', 0), ('JS QS 9H TS KH', 0), ('7C 7S KH 2H 7H', 3), ('3C KH 5D 5S KH', 2), ('QH 8H KD JH 8S', 1), ('2D 6D 9D TH 7D', 0), ) _snake_case = ( ('JH AH TH KH QH', 23), ('JH 9H TH KH QH', 22), ('JC KH JS JD JH', 21), ('KH KC 3S 3H 3D', 20), ('8C 9C 5C 3C TC', 19), ('JS QS 9H TS KH', 18), ('7C 7S KH 2H 7H', 17), ('3C KH 5D 5S KH', 16), ('QH 8H KD JH 8S', 15), ('2D 6D 9D TH 7D', 14), ) def lowerCAmelCase__ ( ): '''simple docstring''' _a , _a : Any = randrange(len(UpperCamelCase__ ) ), randrange(len(UpperCamelCase__ ) ) _a : Any = ["""Loss""", """Tie""", """Win"""][(play >= oppo) + (play > oppo)] _a , _a : List[str] = SORTED_HANDS[play], SORTED_HANDS[oppo] return hand, other, expected def lowerCAmelCase__ ( UpperCamelCase__ = 1_0_0 ): '''simple docstring''' return (generate_random_hand() for _ in range(UpperCamelCase__ )) @pytest.mark.parametrize("""hand, expected""" , UpperCamelCase__ ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' assert PokerHand(UpperCamelCase__ )._is_flush() == expected @pytest.mark.parametrize("""hand, expected""" , UpperCamelCase__ ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' assert PokerHand(UpperCamelCase__ )._is_straight() == expected @pytest.mark.parametrize("""hand, expected, card_values""" , UpperCamelCase__ ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : List[Any] = PokerHand(UpperCamelCase__ ) assert player._is_five_high_straight() == expected assert player._card_values == card_values @pytest.mark.parametrize("""hand, expected""" , UpperCamelCase__ ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' assert PokerHand(UpperCamelCase__ )._is_same_kind() == expected @pytest.mark.parametrize("""hand, expected""" , UpperCamelCase__ ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' assert PokerHand(UpperCamelCase__ )._hand_type == expected @pytest.mark.parametrize("""hand, other, expected""" , UpperCamelCase__ ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' assert PokerHand(UpperCamelCase__ ).compare_with(PokerHand(UpperCamelCase__ ) ) == expected @pytest.mark.parametrize("""hand, other, expected""" , generate_random_hands() ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' assert PokerHand(UpperCamelCase__ ).compare_with(PokerHand(UpperCamelCase__ ) ) == expected def lowerCAmelCase__ ( ): '''simple docstring''' _a : Dict = [PokerHand(UpperCamelCase__ ) for hand in SORTED_HANDS] _a : Tuple = poker_hands.copy() shuffle(UpperCamelCase__ ) _a : Union[str, Any] = chain(sorted(UpperCamelCase__ ) ) for index, hand in enumerate(UpperCamelCase__ ): assert hand == poker_hands[index] def lowerCAmelCase__ ( ): '''simple docstring''' # Test that five high straights are compared correctly. _a : Optional[Any] = [PokerHand("""2D AC 3H 4H 5S""" ), PokerHand("""2S 3H 4H 5S 6C""" )] pokerhands.sort(reverse=UpperCamelCase__ ) assert pokerhands[0].__str__() == "2S 3H 4H 5S 6C" def lowerCAmelCase__ ( ): '''simple docstring''' # Multiple calls to five_high_straight function should still return True # and shouldn't mutate the list in every call other than the first. _a : List[Any] = PokerHand("""2C 4S AS 3D 5C""" ) _a : List[Any] = True _a : List[Any] = [5, 4, 3, 2, 1_4] for _ in range(1_0 ): assert pokerhand._is_five_high_straight() == expected assert pokerhand._card_values == expected_card_values def lowerCAmelCase__ ( ): '''simple docstring''' # Problem number 54 from Project Euler # Testing from poker_hands.txt file _a : str = 0 _a : List[Any] = os.path.abspath(os.path.dirname(UpperCamelCase__ ) ) _a : Any = os.path.join(UpperCamelCase__ , """poker_hands.txt""" ) with open(UpperCamelCase__ ) as file_hand: for line in file_hand: _a : Optional[int] = line[:1_4].strip() _a : Dict = line[1_5:].strip() _a , _a : int = PokerHand(UpperCamelCase__ ), PokerHand(UpperCamelCase__ ) _a : Optional[int] = player.compare_with(UpperCamelCase__ ) if output == "Win": answer += 1 assert answer == 3_7_6
294
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _snake_case = { 'configuration_transfo_xl': ['TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TransfoXLConfig'], 'tokenization_transfo_xl': ['TransfoXLCorpus', 'TransfoXLTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST', 'AdaptiveEmbedding', 'TransfoXLForSequenceClassification', 'TransfoXLLMHeadModel', 'TransfoXLModel', 'TransfoXLPreTrainedModel', 'load_tf_weights_in_transfo_xl', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFAdaptiveEmbedding', 'TFTransfoXLForSequenceClassification', 'TFTransfoXLLMHeadModel', 'TFTransfoXLMainLayer', 'TFTransfoXLModel', 'TFTransfoXLPreTrainedModel', ] if TYPE_CHECKING: from .configuration_transfo_xl import TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, TransfoXLConfig from .tokenization_transfo_xl import TransfoXLCorpus, TransfoXLTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_transfo_xl import ( TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, AdaptiveEmbedding, TransfoXLForSequenceClassification, TransfoXLLMHeadModel, TransfoXLModel, TransfoXLPreTrainedModel, load_tf_weights_in_transfo_xl, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_transfo_xl import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFAdaptiveEmbedding, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLMainLayer, TFTransfoXLModel, TFTransfoXLPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
294
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_clip import CLIPImageProcessor _snake_case = logging.get_logger(__name__) class UpperCamelCase ( snake_case_ ): def __init__( self : Any , *UpperCAmelCase__ : Optional[Any] , **UpperCAmelCase__ : Tuple ) -> None: warnings.warn( """The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use CLIPImageProcessor instead.""" , UpperCAmelCase__ , ) super().__init__(*UpperCAmelCase__ , **UpperCAmelCase__ )
294
"""simple docstring""" def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , ): '''simple docstring''' _a : Optional[Any] = [redshift, radiation_density, matter_density, dark_energy] if any(p < 0 for p in parameters ): raise ValueError("""All input parameters must be positive""" ) if any(p > 1 for p in parameters[1:4] ): raise ValueError("""Relative densities cannot be greater than one""" ) else: _a : Tuple = 1 - (matter_density + radiation_density + dark_energy) _a : int = ( radiation_density * (redshift + 1) ** 4 + matter_density * (redshift + 1) ** 3 + curvature * (redshift + 1) ** 2 + dark_energy ) _a : List[str] = hubble_constant * e_a ** (1 / 2) return hubble if __name__ == "__main__": import doctest # run doctest doctest.testmod() # demo LCDM approximation _snake_case = 0.3 print( hubble_parameter( hubble_constant=68.3, radiation_density=1e-4, matter_density=matter_density, dark_energy=1 - matter_density, redshift=0, ) )
294
1
"""simple docstring""" # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _snake_case = {'configuration_mra': ['MRA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MraConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'MRA_PRETRAINED_MODEL_ARCHIVE_LIST', 'MraForMaskedLM', 'MraForMultipleChoice', 'MraForQuestionAnswering', 'MraForSequenceClassification', 'MraForTokenClassification', 'MraLayer', 'MraModel', 'MraPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mra import MRA_PRETRAINED_CONFIG_ARCHIVE_MAP, MraConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mra import ( MRA_PRETRAINED_MODEL_ARCHIVE_LIST, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraLayer, MraModel, MraPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()['__file__'], _import_structure)
294
"""simple docstring""" import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : List[Any] ) -> Dict: _a : Optional[int] = tempfile.mkdtemp() _a : Optional[Any] = SamImageProcessor() _a : int = SamProcessor(UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) def _lowercase ( self : Tuple , **UpperCAmelCase__ : Any ) -> Any: return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase__ ).image_processor def _lowercase ( self : str ) -> int: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Tuple ) -> Dict: _a : List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _a : Optional[int] = [Image.fromarray(np.moveaxis(UpperCAmelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def _lowercase ( self : Dict ) -> Dict: _a : List[Any] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _a : Tuple = self.get_image_processor(do_normalize=UpperCAmelCase__ , padding_value=1.0 ) _a : Tuple = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=UpperCAmelCase__ , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCAmelCase__ ) def _lowercase ( self : Union[str, Any] ) -> Tuple: _a : Optional[Any] = self.get_image_processor() _a : int = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Union[str, Any] = self.prepare_image_inputs() _a : List[str] = image_processor(UpperCAmelCase__ , return_tensors="""np""" ) _a : List[str] = processor(images=UpperCAmelCase__ , return_tensors="""np""" ) input_feat_extract.pop("""original_sizes""" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("""reshaped_input_sizes""" ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_torch def _lowercase ( self : Optional[Any] ) -> Optional[Any]: _a : Optional[Any] = self.get_image_processor() _a : Dict = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Tuple = [torch.ones((1, 3, 5, 5) )] _a : Tuple = [[1764, 2646]] _a : Optional[int] = [[683, 1024]] _a : List[Any] = processor.post_process_masks(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) _a : int = processor.post_process_masks( UpperCAmelCase__ , torch.tensor(UpperCAmelCase__ ) , torch.tensor(UpperCAmelCase__ ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np _a : Optional[Any] = [np.ones((1, 3, 5, 5) )] _a : Tuple = processor.post_process_masks(UpperCAmelCase__ , np.array(UpperCAmelCase__ ) , np.array(UpperCAmelCase__ ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) _a : List[str] = [[1, 0], [0, 1]] with self.assertRaises(UpperCAmelCase__ ): _a : str = processor.post_process_masks(UpperCAmelCase__ , np.array(UpperCAmelCase__ ) , np.array(UpperCAmelCase__ ) ) @require_vision @require_tf class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : Any ) -> List[str]: _a : List[str] = tempfile.mkdtemp() _a : Any = SamImageProcessor() _a : Union[str, Any] = SamProcessor(UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) def _lowercase ( self : List[str] , **UpperCAmelCase__ : Any ) -> List[str]: return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase__ ).image_processor def _lowercase ( self : Optional[Any] ) -> Union[str, Any]: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Dict ) -> List[str]: _a : List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _a : List[Any] = [Image.fromarray(np.moveaxis(UpperCAmelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def _lowercase ( self : Union[str, Any] ) -> Union[str, Any]: _a : Optional[int] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _a : str = self.get_image_processor(do_normalize=UpperCAmelCase__ , padding_value=1.0 ) _a : Union[str, Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=UpperCAmelCase__ , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCAmelCase__ ) def _lowercase ( self : List[Any] ) -> str: _a : Union[str, Any] = self.get_image_processor() _a : Dict = SamProcessor(image_processor=UpperCAmelCase__ ) _a : int = self.prepare_image_inputs() _a : List[str] = image_processor(UpperCAmelCase__ , return_tensors="""np""" ) _a : List[str] = processor(images=UpperCAmelCase__ , return_tensors="""np""" ) input_feat_extract.pop("""original_sizes""" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("""reshaped_input_sizes""" ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_tf def _lowercase ( self : Optional[Any] ) -> int: _a : Optional[Any] = self.get_image_processor() _a : Dict = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Any = [tf.ones((1, 3, 5, 5) )] _a : Tuple = [[1764, 2646]] _a : str = [[683, 1024]] _a : Union[str, Any] = processor.post_process_masks(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , return_tensors="""tf""" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) _a : Union[str, Any] = processor.post_process_masks( UpperCAmelCase__ , tf.convert_to_tensor(UpperCAmelCase__ ) , tf.convert_to_tensor(UpperCAmelCase__ ) , return_tensors="""tf""" , ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np _a : List[Any] = [np.ones((1, 3, 5, 5) )] _a : Optional[int] = processor.post_process_masks( UpperCAmelCase__ , np.array(UpperCAmelCase__ ) , np.array(UpperCAmelCase__ ) , return_tensors="""tf""" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) _a : Dict = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): _a : List[Any] = processor.post_process_masks( UpperCAmelCase__ , np.array(UpperCAmelCase__ ) , np.array(UpperCAmelCase__ ) , return_tensors="""tf""" ) @require_vision @require_torchvision class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : str ) -> Optional[Any]: _a : Optional[Any] = tempfile.mkdtemp() _a : Dict = SamImageProcessor() _a : List[str] = SamProcessor(UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) def _lowercase ( self : Any , **UpperCAmelCase__ : Dict ) -> int: return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase__ ).image_processor def _lowercase ( self : Tuple ) -> List[Any]: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : str ) -> int: _a : str = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _a : int = [Image.fromarray(np.moveaxis(UpperCAmelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def _lowercase ( self : int ) -> List[Any]: _a : Optional[Any] = self.get_image_processor() _a : Optional[Any] = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Tuple = np.random.randint(0 , 2 , size=(1, 3, 5, 5) ).astype(np.floataa ) _a : str = [tf.convert_to_tensor(UpperCAmelCase__ )] _a : Optional[int] = [torch.tensor(UpperCAmelCase__ )] _a : Union[str, Any] = [[1764, 2646]] _a : List[str] = [[683, 1024]] _a : Optional[int] = processor.post_process_masks( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , return_tensors="""tf""" ) _a : List[str] = processor.post_process_masks( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , return_tensors="""pt""" ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def _lowercase ( self : str ) -> Optional[Any]: _a : List[Any] = self.get_image_processor() _a : Any = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Dict = self.prepare_image_inputs() _a : List[str] = image_processor(UpperCAmelCase__ , return_tensors="""pt""" )["""pixel_values"""].numpy() _a : str = processor(images=UpperCAmelCase__ , return_tensors="""pt""" )["""pixel_values"""].numpy() _a : Optional[Any] = image_processor(UpperCAmelCase__ , return_tensors="""tf""" )["""pixel_values"""].numpy() _a : Optional[int] = processor(images=UpperCAmelCase__ , return_tensors="""tf""" )["""pixel_values"""].numpy() self.assertTrue(np.allclose(UpperCAmelCase__ , UpperCAmelCase__ ) ) self.assertTrue(np.allclose(UpperCAmelCase__ , UpperCAmelCase__ ) ) self.assertTrue(np.allclose(UpperCAmelCase__ , UpperCAmelCase__ ) )
294
1
"""simple docstring""" import pytest from datasets import inspect_metric, list_metrics, load_metric @pytest.fixture def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' monkeypatch.setattr("""datasets.utils.deprecation_utils._emitted_deprecation_warnings""" , set() ) @pytest.fixture def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' class UpperCamelCase : def __init__( self : List[str] , UpperCAmelCase__ : List[Any] ) -> List[str]: _a : Dict = metric_id class UpperCamelCase : UpperCamelCase : Dict = [MetricMock(snake_case_ ) for metric_id in ['''accuracy''', '''mse''', '''precision''', '''codeparrot/apps_metric''']] def _lowercase ( self : Union[str, Any] ) -> Optional[Any]: return self._metrics monkeypatch.setattr("""datasets.inspect.huggingface_hub""" , HfhMock() ) @pytest.mark.parametrize( """func, args""" , [(load_metric, ("""metrics/mse""",)), (list_metrics, ()), (inspect_metric, ("""metrics/mse""", """tmp_path"""))] ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' if "tmp_path" in args: _a : Tuple = tuple(arg if arg != """tmp_path""" else tmp_path for arg in args ) with pytest.warns(UpperCamelCase__ , match="""https://huggingface.co/docs/evaluate""" ): func(*UpperCamelCase__ )
294
"""simple docstring""" import argparse import gc import json import os import re import torch from huggingface_hub import hf_hub_download from transformers import AutoModelForCausalLM, AutoTokenizer, PreTrainedTokenizerFast, RwkvConfig from transformers.modeling_utils import WEIGHTS_INDEX_NAME, shard_checkpoint _snake_case = { '169M': 12, '430M': 24, '1B5': 24, '3B': 32, '7B': 32, '14B': 40, } _snake_case = { '169M': 768, '430M': 1024, '1B5': 2048, '3B': 2560, '7B': 4096, '14B': 5120, } def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : int = list(state_dict.keys() ) for name in state_dict_keys: _a : str = state_dict.pop(UpperCamelCase__ ) # emb -> embedding if name.startswith("""emb.""" ): _a : Dict = name.replace("""emb.""" , """embeddings.""" ) # ln_0 -> pre_ln (only present at block 0) if name.startswith("""blocks.0.ln0""" ): _a : Dict = name.replace("""blocks.0.ln0""" , """blocks.0.pre_ln""" ) # att -> attention _a : Any = re.sub(R"""blocks\.(\d+)\.att""" , R"""blocks.\1.attention""" , UpperCamelCase__ ) # ffn -> feed_forward _a : int = re.sub(R"""blocks\.(\d+)\.ffn""" , R"""blocks.\1.feed_forward""" , UpperCamelCase__ ) # time_mix_k -> time_mix_key and reshape if name.endswith(""".time_mix_k""" ): _a : List[str] = name.replace(""".time_mix_k""" , """.time_mix_key""" ) # time_mix_v -> time_mix_value and reshape if name.endswith(""".time_mix_v""" ): _a : Tuple = name.replace(""".time_mix_v""" , """.time_mix_value""" ) # time_mix_r -> time_mix_key and reshape if name.endswith(""".time_mix_r""" ): _a : Dict = name.replace(""".time_mix_r""" , """.time_mix_receptance""" ) if name != "head.weight": _a : Optional[int] = """rwkv.""" + name _a : Any = weight return state_dict def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=False , UpperCamelCase__=None ): '''simple docstring''' # 1. If possible, build the tokenizer. if tokenizer_file is None: print("""No `--tokenizer_file` provided, we will use the default tokenizer.""" ) _a : Tuple = 5_0_2_7_7 _a : str = AutoTokenizer.from_pretrained("""EleutherAI/gpt-neox-20b""" ) else: _a : int = PreTrainedTokenizerFast(tokenizer_file=UpperCamelCase__ ) _a : int = len(UpperCamelCase__ ) tokenizer.save_pretrained(UpperCamelCase__ ) # 2. Build the config _a : Optional[Any] = list(NUM_HIDDEN_LAYERS_MAPPING.keys() ) if size is None: # Try to infer size from the checkpoint name for candidate in possible_sizes: if candidate in checkpoint_file: _a : Tuple = candidate break if size is None: raise ValueError("""Could not infer the size, please provide it with the `--size` argument.""" ) if size not in possible_sizes: raise ValueError(F"""`size` should be one of {possible_sizes}, got {size}.""" ) _a : List[Any] = RwkvConfig( vocab_size=UpperCamelCase__ , num_hidden_layers=NUM_HIDDEN_LAYERS_MAPPING[size] , hidden_size=HIDEN_SIZE_MAPPING[size] , ) config.save_pretrained(UpperCamelCase__ ) # 3. Download model file then convert state_dict _a : str = hf_hub_download(UpperCamelCase__ , UpperCamelCase__ ) _a : int = torch.load(UpperCamelCase__ , map_location="""cpu""" ) _a : List[str] = convert_state_dict(UpperCamelCase__ ) # 4. Split in shards and save _a , _a : List[str] = shard_checkpoint(UpperCamelCase__ ) for shard_file, shard in shards.items(): torch.save(UpperCamelCase__ , os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) if index is not None: _a : Optional[Any] = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) # Save the index as well with open(UpperCamelCase__ , """w""" , encoding="""utf-8""" ) as f: _a : Dict = json.dumps(UpperCamelCase__ , indent=2 , sort_keys=UpperCamelCase__ ) + """\n""" f.write(UpperCamelCase__ ) # 5. Clean up shards (for some reason the file PyTorch saves take the same space as the whole state_dict print( """Cleaning up shards. This may error with an OOM error, it this is the case don't worry you still have converted the model.""" ) _a : List[Any] = list(shards.keys() ) del state_dict del shards gc.collect() for shard_file in shard_files: _a : Any = torch.load(os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) torch.save({k: v.cpu().clone() for k, v in state_dict.items()} , os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) del state_dict gc.collect() if push_to_hub: if model_name is None: raise ValueError("""Please provide a `model_name` to push the model to the Hub.""" ) _a : Dict = AutoModelForCausalLM.from_pretrained(UpperCamelCase__ ) model.push_to_hub(UpperCamelCase__ , max_shard_size="""2GB""" ) tokenizer.push_to_hub(UpperCamelCase__ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '--repo_id', default=None, type=str, required=True, help='Repo ID from which to pull the checkpoint.' ) parser.add_argument( '--checkpoint_file', default=None, type=str, required=True, help='Name of the checkpoint file in the repo.' ) parser.add_argument( '--output_dir', default=None, type=str, required=True, help='Where to save the converted model.' ) parser.add_argument( '--tokenizer_file', default=None, type=str, help='Path to the tokenizer file to use (if not provided, only the model is converted).', ) parser.add_argument( '--size', default=None, type=str, help='Size of the model. Will be inferred from the `checkpoint_file` if not passed.', ) parser.add_argument( '--push_to_hub', action='store_true', help='Push to the Hub the converted model.', ) parser.add_argument( '--model_name', default=None, type=str, help='Name of the pushed model on the Hub, including the username / organization.', ) _snake_case = parser.parse_args() convert_rmkv_checkpoint_to_hf_format( args.repo_id, args.checkpoint_file, args.output_dir, size=args.size, tokenizer_file=args.tokenizer_file, push_to_hub=args.push_to_hub, model_name=args.model_name, )
294
1
"""simple docstring""" import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import YolosImageProcessor class UpperCamelCase ( unittest.TestCase ): def __init__( self : Tuple , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : str=7 , UpperCAmelCase__ : Tuple=3 , UpperCAmelCase__ : List[str]=30 , UpperCAmelCase__ : Optional[int]=400 , UpperCAmelCase__ : Union[str, Any]=True , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : Optional[Any]=True , UpperCAmelCase__ : Union[str, Any]=[0.5, 0.5, 0.5] , UpperCAmelCase__ : List[str]=[0.5, 0.5, 0.5] , UpperCAmelCase__ : List[str]=True , UpperCAmelCase__ : Tuple=1 / 255 , UpperCAmelCase__ : Union[str, Any]=True , ) -> Tuple: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p _a : Optional[Any] = size if size is not None else {"""shortest_edge""": 18, """longest_edge""": 1333} _a : Optional[Any] = parent _a : List[str] = batch_size _a : Optional[Any] = num_channels _a : Optional[Any] = min_resolution _a : str = max_resolution _a : Tuple = do_resize _a : Any = size _a : List[Any] = do_normalize _a : List[Any] = image_mean _a : Union[str, Any] = image_std _a : int = do_rescale _a : Optional[int] = rescale_factor _a : List[Any] = do_pad def _lowercase ( self : Optional[Any] ) -> Dict: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def _lowercase ( self : Optional[int] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : int=False ) -> int: if not batched: _a : Tuple = image_inputs[0] if isinstance(UpperCAmelCase__ , Image.Image ): _a , _a : Union[str, Any] = image.size else: _a , _a : Tuple = image.shape[1], image.shape[2] if w < h: _a : Any = int(self.size["""shortest_edge"""] * h / w ) _a : Optional[int] = self.size["""shortest_edge"""] elif w > h: _a : Optional[Any] = self.size["""shortest_edge"""] _a : List[Any] = int(self.size["""shortest_edge"""] * w / h ) else: _a : List[Any] = self.size["""shortest_edge"""] _a : str = self.size["""shortest_edge"""] else: _a : int = [] for image in image_inputs: _a , _a : Any = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) _a : Dict = max(UpperCAmelCase__ , key=lambda UpperCAmelCase__ : item[0] )[0] _a : Optional[int] = max(UpperCAmelCase__ , key=lambda UpperCAmelCase__ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class UpperCamelCase ( snake_case_ , unittest.TestCase ): UpperCamelCase : int = YolosImageProcessor if is_vision_available() else None def _lowercase ( self : Dict ) -> Union[str, Any]: _a : int = YolosImageProcessingTester(self ) @property def _lowercase ( self : Any ) -> List[Any]: return self.image_processor_tester.prepare_image_processor_dict() def _lowercase ( self : Optional[Any] ) -> Optional[Any]: _a : Any = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCAmelCase__ , """image_mean""" ) ) self.assertTrue(hasattr(UpperCAmelCase__ , """image_std""" ) ) self.assertTrue(hasattr(UpperCAmelCase__ , """do_normalize""" ) ) self.assertTrue(hasattr(UpperCAmelCase__ , """do_resize""" ) ) self.assertTrue(hasattr(UpperCAmelCase__ , """size""" ) ) def _lowercase ( self : Tuple ) -> List[Any]: _a : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""shortest_edge""": 18, """longest_edge""": 1333} ) self.assertEqual(image_processor.do_pad , UpperCAmelCase__ ) _a : Optional[int] = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=UpperCAmelCase__ ) self.assertEqual(image_processor.size , {"""shortest_edge""": 42, """longest_edge""": 84} ) self.assertEqual(image_processor.do_pad , UpperCAmelCase__ ) def _lowercase ( self : int ) -> Optional[int]: pass def _lowercase ( self : Union[str, Any] ) -> Dict: # Initialize image_processing _a : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _a : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , Image.Image ) # Test not batched input _a : str = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values _a , _a : Dict = self.image_processor_tester.get_expected_values(UpperCAmelCase__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _a , _a : Union[str, Any] = self.image_processor_tester.get_expected_values(UpperCAmelCase__ , batched=UpperCAmelCase__ ) _a : List[str] = image_processing(UpperCAmelCase__ , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _lowercase ( self : Optional[Any] ) -> Tuple: # Initialize image_processing _a : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _a : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ , numpify=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , np.ndarray ) # Test not batched input _a : Optional[int] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values _a , _a : Optional[int] = self.image_processor_tester.get_expected_values(UpperCAmelCase__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _a : int = image_processing(UpperCAmelCase__ , return_tensors="""pt""" ).pixel_values _a , _a : Tuple = self.image_processor_tester.get_expected_values(UpperCAmelCase__ , batched=UpperCAmelCase__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _lowercase ( self : List[str] ) -> Dict: # Initialize image_processing _a : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _a : List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ , torchify=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , torch.Tensor ) # Test not batched input _a : Optional[int] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values _a , _a : Any = self.image_processor_tester.get_expected_values(UpperCAmelCase__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _a : str = image_processing(UpperCAmelCase__ , return_tensors="""pt""" ).pixel_values _a , _a : Dict = self.image_processor_tester.get_expected_values(UpperCAmelCase__ , batched=UpperCAmelCase__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _lowercase ( self : Union[str, Any] ) -> Optional[Any]: # Initialize image_processings _a : List[Any] = self.image_processing_class(**self.image_processor_dict ) _a : Any = self.image_processing_class(do_resize=UpperCAmelCase__ , do_normalize=UpperCAmelCase__ , do_rescale=UpperCAmelCase__ ) # create random PyTorch tensors _a : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCAmelCase__ , torchify=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__ , torch.Tensor ) # Test whether the method "pad" and calling the image processor return the same tensors _a : Optional[Any] = image_processing_a.pad(UpperCAmelCase__ , return_tensors="""pt""" ) _a : str = image_processing_a(UpperCAmelCase__ , return_tensors="""pt""" ) self.assertTrue( torch.allclose(encoded_images_with_method["""pixel_values"""] , encoded_images["""pixel_values"""] , atol=1E-4 ) ) @slow def _lowercase ( self : Optional[Any] ) -> Tuple: # prepare image and target _a : int = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) with open("""./tests/fixtures/tests_samples/COCO/coco_annotations.txt""" , """r""" ) as f: _a : List[str] = json.loads(f.read() ) _a : Optional[int] = {"""image_id""": 39769, """annotations""": target} # encode them _a : Any = YolosImageProcessor.from_pretrained("""hustvl/yolos-small""" ) _a : List[Any] = image_processing(images=UpperCAmelCase__ , annotations=UpperCAmelCase__ , return_tensors="""pt""" ) # verify pixel values _a : Any = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["""pixel_values"""].shape , UpperCAmelCase__ ) _a : Optional[int] = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding["""pixel_values"""][0, 0, 0, :3] , UpperCAmelCase__ , atol=1E-4 ) ) # verify area _a : int = torch.tensor([5_8_8_7.9_6_0_0, 1_1_2_5_0.2_0_6_1, 4_8_9_3_5_3.8_4_3_8, 8_3_7_1_2_2.7_5_0_0, 1_4_7_9_6_7.5_1_5_6, 1_6_5_7_3_2.3_4_3_8] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""area"""] , UpperCAmelCase__ ) ) # verify boxes _a : Dict = torch.Size([6, 4] ) self.assertEqual(encoding["""labels"""][0]["""boxes"""].shape , UpperCAmelCase__ ) _a : List[Any] = torch.tensor([0.5_5_0_3, 0.2_7_6_5, 0.0_6_0_4, 0.2_2_1_5] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""boxes"""][0] , UpperCAmelCase__ , atol=1E-3 ) ) # verify image_id _a : Optional[int] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""image_id"""] , UpperCAmelCase__ ) ) # verify is_crowd _a : Optional[Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""iscrowd"""] , UpperCAmelCase__ ) ) # verify class_labels _a : Optional[Any] = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""class_labels"""] , UpperCAmelCase__ ) ) # verify orig_size _a : Optional[int] = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""orig_size"""] , UpperCAmelCase__ ) ) # verify size _a : List[Any] = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""size"""] , UpperCAmelCase__ ) ) @slow def _lowercase ( self : Union[str, Any] ) -> Dict: # prepare image, target and masks_path _a : Dict = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) with open("""./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt""" , """r""" ) as f: _a : Optional[Any] = json.loads(f.read() ) _a : Optional[Any] = {"""file_name""": """000000039769.png""", """image_id""": 39769, """segments_info""": target} _a : List[Any] = pathlib.Path("""./tests/fixtures/tests_samples/COCO/coco_panoptic""" ) # encode them _a : Dict = YolosImageProcessor(format="""coco_panoptic""" ) _a : Union[str, Any] = image_processing(images=UpperCAmelCase__ , annotations=UpperCAmelCase__ , masks_path=UpperCAmelCase__ , return_tensors="""pt""" ) # verify pixel values _a : str = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["""pixel_values"""].shape , UpperCAmelCase__ ) _a : Tuple = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding["""pixel_values"""][0, 0, 0, :3] , UpperCAmelCase__ , atol=1E-4 ) ) # verify area _a : Union[str, Any] = torch.tensor([1_4_7_9_7_9.6_8_7_5, 1_6_5_5_2_7.0_4_6_9, 4_8_4_6_3_8.5_9_3_8, 1_1_2_9_2.9_3_7_5, 5_8_7_9.6_5_6_2, 7_6_3_4.1_1_4_7] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""area"""] , UpperCAmelCase__ ) ) # verify boxes _a : List[str] = torch.Size([6, 4] ) self.assertEqual(encoding["""labels"""][0]["""boxes"""].shape , UpperCAmelCase__ ) _a : List[str] = torch.tensor([0.2_6_2_5, 0.5_4_3_7, 0.4_6_8_8, 0.8_6_2_5] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""boxes"""][0] , UpperCAmelCase__ , atol=1E-3 ) ) # verify image_id _a : str = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""image_id"""] , UpperCAmelCase__ ) ) # verify is_crowd _a : List[str] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""iscrowd"""] , UpperCAmelCase__ ) ) # verify class_labels _a : Any = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""class_labels"""] , UpperCAmelCase__ ) ) # verify masks _a : int = 822873 self.assertEqual(encoding["""labels"""][0]["""masks"""].sum().item() , UpperCAmelCase__ ) # verify orig_size _a : Any = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""orig_size"""] , UpperCAmelCase__ ) ) # verify size _a : Any = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""size"""] , UpperCAmelCase__ ) )
294
"""simple docstring""" import shutil import tempfile import unittest from transformers import ClapFeatureExtractor, ClapProcessor, RobertaTokenizer, RobertaTokenizerFast from transformers.testing_utils import require_sentencepiece, require_torchaudio from .test_feature_extraction_clap import floats_list @require_torchaudio @require_sentencepiece class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : int ) -> List[str]: _a : Any = """laion/clap-htsat-unfused""" _a : Union[str, Any] = tempfile.mkdtemp() def _lowercase ( self : List[Any] , **UpperCAmelCase__ : Any ) -> Dict: return RobertaTokenizer.from_pretrained(self.checkpoint , **UpperCAmelCase__ ) def _lowercase ( self : List[Any] , **UpperCAmelCase__ : List[str] ) -> int: return ClapFeatureExtractor.from_pretrained(self.checkpoint , **UpperCAmelCase__ ) def _lowercase ( self : List[Any] ) -> Tuple: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : List[str] ) -> Optional[int]: _a : List[str] = self.get_tokenizer() _a : Any = self.get_feature_extractor() _a : Optional[Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) _a : List[str] = ClapProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCAmelCase__ ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , UpperCAmelCase__ ) def _lowercase ( self : Dict ) -> Optional[int]: _a : Tuple = ClapProcessor(tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() ) processor.save_pretrained(self.tmpdirname ) _a : Dict = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) _a : Union[str, Any] = self.get_feature_extractor(do_normalize=UpperCAmelCase__ , padding_value=1.0 ) _a : Union[str, Any] = ClapProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=UpperCAmelCase__ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCAmelCase__ ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.feature_extractor , UpperCAmelCase__ ) def _lowercase ( self : List[str] ) -> Optional[Any]: _a : Optional[int] = self.get_feature_extractor() _a : Tuple = self.get_tokenizer() _a : List[Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) _a : Any = floats_list((3, 1000) ) _a : List[Any] = feature_extractor(UpperCAmelCase__ , return_tensors="""np""" ) _a : List[str] = processor(audios=UpperCAmelCase__ , return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def _lowercase ( self : Tuple ) -> Optional[int]: _a : List[str] = self.get_feature_extractor() _a : Any = self.get_tokenizer() _a : Any = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) _a : Optional[int] = """This is a test string""" _a : Tuple = processor(text=UpperCAmelCase__ ) _a : int = tokenizer(UpperCAmelCase__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _lowercase ( self : List[Any] ) -> Any: _a : str = self.get_feature_extractor() _a : List[str] = self.get_tokenizer() _a : List[Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) _a : Any = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _a : Dict = processor.batch_decode(UpperCAmelCase__ ) _a : Any = tokenizer.batch_decode(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def _lowercase ( self : Dict ) -> List[str]: _a : str = self.get_feature_extractor() _a : Optional[Any] = self.get_tokenizer() _a : Union[str, Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) self.assertListEqual( processor.model_input_names[2:] , feature_extractor.model_input_names , msg="""`processor` and `feature_extractor` model input names do not match""" , )
294
1
"""simple docstring""" import pytest _snake_case = '__dummy_dataset1__' _snake_case = '\nimport json\nimport os\n\nimport datasets\n\n\nREPO_URL = "https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/"\nURLS = {"train": REPO_URL + "wikiann-bn-train.jsonl", "validation": REPO_URL + "wikiann-bn-validation.jsonl"}\n\n\nclass __DummyDataset1__(datasets.GeneratorBasedBuilder):\n\n def _info(self):\n features = datasets.Features(\n {\n "tokens": datasets.Sequence(datasets.Value("string")),\n "ner_tags": datasets.Sequence(\n datasets.features.ClassLabel(\n names=[\n "O",\n "B-PER",\n "I-PER",\n "B-ORG",\n "I-ORG",\n "B-LOC",\n "I-LOC",\n ]\n )\n ),\n "langs": datasets.Sequence(datasets.Value("string")),\n "spans": datasets.Sequence(datasets.Value("string")),\n }\n )\n return datasets.DatasetInfo(features=features)\n\n def _split_generators(self, dl_manager):\n dl_path = dl_manager.download(URLS)\n return [\n datasets.SplitGenerator(datasets.Split.TRAIN, gen_kwargs={"filepath": dl_path["train"]}),\n datasets.SplitGenerator(datasets.Split.VALIDATION, gen_kwargs={"filepath": dl_path["validation"]}),\n ]\n\n def _generate_examples(self, filepath):\n with open(filepath, "r", encoding="utf-8") as f:\n for i, line in enumerate(f):\n yield i, json.loads(line)\n' @pytest.fixture def lowerCAmelCase__ ( ): '''simple docstring''' return DATASET_LOADING_SCRIPT_NAME @pytest.fixture def lowerCAmelCase__ ( ): '''simple docstring''' return DATASET_LOADING_SCRIPT_CODE @pytest.fixture def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Optional[int] = dataset_loading_script_name _a : List[Any] = tmp_path / """datasets""" / script_name script_dir.mkdir(parents=UpperCamelCase__ ) _a : Optional[int] = script_dir / F"""{script_name}.py""" with open(UpperCamelCase__ , """w""" ) as f: f.write(UpperCamelCase__ ) return str(UpperCamelCase__ )
294
"""simple docstring""" import warnings from ...utils import logging from .image_processing_clip import CLIPImageProcessor _snake_case = logging.get_logger(__name__) class UpperCamelCase ( snake_case_ ): def __init__( self : Any , *UpperCAmelCase__ : Optional[Any] , **UpperCAmelCase__ : Tuple ) -> None: warnings.warn( """The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use CLIPImageProcessor instead.""" , UpperCAmelCase__ , ) super().__init__(*UpperCAmelCase__ , **UpperCAmelCase__ )
294
1
"""simple docstring""" import argparse import shutil from pathlib import Path from tqdm import tqdm from transformers import AutoTokenizer def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=1_0_2_4 ): '''simple docstring''' _a , _a : Any = [], [] _a : str = list(zip(UpperCamelCase__ , UpperCamelCase__ ) ) _a , _a : Tuple = sorted_examples[0] def is_too_big(UpperCamelCase__ ): return tok(UpperCamelCase__ , return_tensors="""pt""" ).input_ids.shape[1] > max_tokens for src, tgt in tqdm(sorted_examples[1:] ): _a : Tuple = new_src + """ """ + src _a : str = new_tgt + """ """ + tgt if is_too_big(UpperCamelCase__ ) or is_too_big(UpperCamelCase__ ): # cant fit, finalize example finished_src.append(UpperCamelCase__ ) finished_tgt.append(UpperCamelCase__ ) _a , _a : List[Any] = src, tgt else: # can fit, keep adding _a , _a : Optional[int] = cand_src, cand_tgt # cleanup if new_src: assert new_tgt finished_src.append(UpperCamelCase__ ) finished_tgt.append(UpperCamelCase__ ) return finished_src, finished_tgt def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : List[str] = Path(UpperCamelCase__ ) save_path.mkdir(exist_ok=UpperCamelCase__ ) for split in ["train"]: _a , _a : Union[str, Any] = data_dir / F"""{split}.source""", data_dir / F"""{split}.target""" _a : Any = [x.rstrip() for x in Path(UpperCamelCase__ ).open().readlines()] _a : Any = [x.rstrip() for x in Path(UpperCamelCase__ ).open().readlines()] _a , _a : Tuple = pack_examples(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) print(F"""packed {split} split from {len(UpperCamelCase__ )} examples -> {len(UpperCamelCase__ )}.""" ) Path(save_path / F"""{split}.source""" ).open("""w""" ).write("""\n""".join(UpperCamelCase__ ) ) Path(save_path / F"""{split}.target""" ).open("""w""" ).write("""\n""".join(UpperCamelCase__ ) ) for split in ["val", "test"]: _a , _a : int = data_dir / F"""{split}.source""", data_dir / F"""{split}.target""" shutil.copyfile(UpperCamelCase__ , save_path / F"""{split}.source""" ) shutil.copyfile(UpperCamelCase__ , save_path / F"""{split}.target""" ) def lowerCAmelCase__ ( ): '''simple docstring''' _a : Union[str, Any] = argparse.ArgumentParser() parser.add_argument("""--tok_name""" , type=UpperCamelCase__ , help="""like facebook/bart-large-cnn,t5-base, etc.""" ) parser.add_argument("""--max_seq_len""" , type=UpperCamelCase__ , default=1_2_8 ) parser.add_argument("""--data_dir""" , type=UpperCamelCase__ ) parser.add_argument("""--save_path""" , type=UpperCamelCase__ ) _a : int = parser.parse_args() _a : int = AutoTokenizer.from_pretrained(args.tok_name ) return pack_data_dir(UpperCamelCase__ , Path(args.data_dir ) , args.max_seq_len , args.save_path ) if __name__ == "__main__": packer_cli()
294
"""simple docstring""" import unittest import numpy as np def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = None , ): '''simple docstring''' _a : List[Any] = np.shape(UpperCamelCase__ ) _a : Any = np.shape(UpperCamelCase__ ) _a : Union[str, Any] = np.shape(UpperCamelCase__ ) if shape_a[0] != shape_b[0]: _a : int = ( """Expected the same number of rows for A and B. """ F"""Instead found A of size {shape_a} and B of size {shape_b}""" ) raise ValueError(UpperCamelCase__ ) if shape_b[1] != shape_c[1]: _a : Tuple = ( """Expected the same number of columns for B and C. """ F"""Instead found B of size {shape_b} and C of size {shape_c}""" ) raise ValueError(UpperCamelCase__ ) _a : int = pseudo_inv if a_inv is None: try: _a : Optional[int] = np.linalg.inv(UpperCamelCase__ ) except np.linalg.LinAlgError: raise ValueError( """Input matrix A is not invertible. Cannot compute Schur complement.""" ) return mat_c - mat_b.T @ a_inv @ mat_b class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : int ) -> None: _a : str = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) _a : Tuple = np.array([[0, 3], [3, 0], [2, 3]] ) _a : Optional[int] = np.array([[2, 1], [6, 3]] ) _a : Optional[Any] = schur_complement(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) _a : Union[str, Any] = np.block([[a, b], [b.T, c]] ) _a : int = np.linalg.det(UpperCAmelCase__ ) _a : Union[str, Any] = np.linalg.det(UpperCAmelCase__ ) _a : List[Any] = np.linalg.det(UpperCAmelCase__ ) self.assertAlmostEqual(UpperCAmelCase__ , det_a * det_s ) def _lowercase ( self : int ) -> None: _a : Optional[int] = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) _a : Optional[int] = np.array([[0, 3], [3, 0], [2, 3]] ) _a : Union[str, Any] = np.array([[2, 1], [6, 3]] ) with self.assertRaises(UpperCAmelCase__ ): schur_complement(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) def _lowercase ( self : List[Any] ) -> None: _a : Any = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) _a : Dict = np.array([[0, 3], [3, 0], [2, 3]] ) _a : List[Any] = np.array([[2, 1, 3], [6, 3, 5]] ) with self.assertRaises(UpperCAmelCase__ ): schur_complement(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) if __name__ == "__main__": import doctest doctest.testmod() unittest.main()
294
1
"""simple docstring""" import gc import unittest from parameterized import parameterized from diffusers import FlaxUNetaDConditionModel from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import load_hf_numpy, require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp @slow @require_flax class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : int , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Optional[Any] ) -> Any: return f"""gaussian_noise_s={seed}_shape={'_'.join([str(UpperCAmelCase__ ) for s in shape] )}.npy""" def _lowercase ( self : Optional[Any] ) -> Optional[int]: # clean up the VRAM after each test super().tearDown() gc.collect() def _lowercase ( self : str , UpperCAmelCase__ : List[Any]=0 , UpperCAmelCase__ : Optional[int]=(4, 4, 64, 64) , UpperCAmelCase__ : Any=False ) -> Any: _a : str = jnp.bfloataa if fpaa else jnp.floataa _a : List[str] = jnp.array(load_hf_numpy(self.get_file_format(UpperCAmelCase__ , UpperCAmelCase__ ) ) , dtype=UpperCAmelCase__ ) return image def _lowercase ( self : Tuple , UpperCAmelCase__ : List[Any]=False , UpperCAmelCase__ : str="CompVis/stable-diffusion-v1-4" ) -> Union[str, Any]: _a : List[Any] = jnp.bfloataa if fpaa else jnp.floataa _a : List[str] = """bf16""" if fpaa else None _a , _a : Tuple = FlaxUNetaDConditionModel.from_pretrained( UpperCAmelCase__ , subfolder="""unet""" , dtype=UpperCAmelCase__ , revision=UpperCAmelCase__ ) return model, params def _lowercase ( self : int , UpperCAmelCase__ : List[Any]=0 , UpperCAmelCase__ : List[str]=(4, 77, 768) , UpperCAmelCase__ : Dict=False ) -> Dict: _a : int = jnp.bfloataa if fpaa else jnp.floataa _a : Dict = jnp.array(load_hf_numpy(self.get_file_format(UpperCAmelCase__ , UpperCAmelCase__ ) ) , dtype=UpperCAmelCase__ ) return hidden_states @parameterized.expand( [ # fmt: off [83, 4, [-0.2_3_2_3, -0.1_3_0_4, 0.0_8_1_3, -0.3_0_9_3, -0.0_9_1_9, -0.1_5_7_1, -0.1_1_2_5, -0.5_8_0_6]], [17, 0.5_5, [-0.0_8_3_1, -0.2_4_4_3, 0.0_9_0_1, -0.0_9_1_9, 0.3_3_9_6, 0.0_1_0_3, -0.3_7_4_3, 0.0_7_0_1]], [8, 0.8_9, [-0.4_8_6_3, 0.0_8_5_9, 0.0_8_7_5, -0.1_6_5_8, 0.9_1_9_9, -0.0_1_1_4, 0.4_8_3_9, 0.4_6_3_9]], [3, 1000, [-0.5_6_4_9, 0.2_4_0_2, -0.5_5_1_8, 0.1_2_4_8, 1.1_3_2_8, -0.2_4_4_3, -0.0_3_2_5, -1.0_0_7_8]], # fmt: on ] ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Any , UpperCAmelCase__ : Optional[Any] ) -> List[str]: _a , _a : str = self.get_unet_model(model_id="""CompVis/stable-diffusion-v1-4""" , fpaa=UpperCAmelCase__ ) _a : int = self.get_latents(UpperCAmelCase__ , fpaa=UpperCAmelCase__ ) _a : Tuple = self.get_encoder_hidden_states(UpperCAmelCase__ , fpaa=UpperCAmelCase__ ) _a : List[str] = model.apply( {"""params""": params} , UpperCAmelCase__ , jnp.array(UpperCAmelCase__ , dtype=jnp.intaa ) , encoder_hidden_states=UpperCAmelCase__ , ).sample assert sample.shape == latents.shape _a : Optional[int] = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) _a : Optional[int] = jnp.array(UpperCAmelCase__ , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, in the same hardware assert jnp.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1E-2 ) @parameterized.expand( [ # fmt: off [83, 4, [0.1_5_1_4, 0.0_8_0_7, 0.1_6_2_4, 0.1_0_1_6, -0.1_8_9_6, 0.0_2_6_3, 0.0_6_7_7, 0.2_3_1_0]], [17, 0.5_5, [0.1_1_6_4, -0.0_2_1_6, 0.0_1_7_0, 0.1_5_8_9, -0.3_1_2_0, 0.1_0_0_5, -0.0_5_8_1, -0.1_4_5_8]], [8, 0.8_9, [-0.1_7_5_8, -0.0_1_6_9, 0.1_0_0_4, -0.1_4_1_1, 0.1_3_1_2, 0.1_1_0_3, -0.1_9_9_6, 0.2_1_3_9]], [3, 1000, [0.1_2_1_4, 0.0_3_5_2, -0.0_7_3_1, -0.1_5_6_2, -0.0_9_9_4, -0.0_9_0_6, -0.2_3_4_0, -0.0_5_3_9]], # fmt: on ] ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Any , UpperCAmelCase__ : str ) -> Any: _a , _a : Dict = self.get_unet_model(model_id="""stabilityai/stable-diffusion-2""" , fpaa=UpperCAmelCase__ ) _a : str = self.get_latents(UpperCAmelCase__ , shape=(4, 4, 96, 96) , fpaa=UpperCAmelCase__ ) _a : Any = self.get_encoder_hidden_states(UpperCAmelCase__ , shape=(4, 77, 1024) , fpaa=UpperCAmelCase__ ) _a : List[str] = model.apply( {"""params""": params} , UpperCAmelCase__ , jnp.array(UpperCAmelCase__ , dtype=jnp.intaa ) , encoder_hidden_states=UpperCAmelCase__ , ).sample assert sample.shape == latents.shape _a : Dict = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) _a : Dict = jnp.array(UpperCAmelCase__ , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, on the same hardware assert jnp.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1E-2 )
294
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _snake_case = logging.get_logger(__name__) _snake_case = torch.device('cpu') def lowerCAmelCase__ ( ): '''simple docstring''' _a : int = """http://images.cocodataset.org/val2017/000000039769.jpg""" _a : Dict = Image.open(requests.get(UpperCamelCase__ , stream=UpperCamelCase__ ).raw ) return im def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' if swiftformer_name == "swiftformer_xs": return torch.tensor([-2.17_03e00, 2.11_07e00, -2.08_11e00, 8.86_85e-01, 2.43_60e-01] ) elif swiftformer_name == "swiftformer_s": return torch.tensor([3.96_36e-01, 2.34_78e-01, -1.69_63e00, -1.73_81e00, -8.63_37e-01] ) elif swiftformer_name == "swiftformer_l1": return torch.tensor([-4.27_68e-01, -4.74_29e-01, -1.08_97e00, -1.02_48e00, 3.55_23e-02] ) elif swiftformer_name == "swiftformer_l3": return torch.tensor([-2.53_30e-01, 2.42_11e-01, -6.01_85e-01, -8.27_89e-01, -6.04_46e-02] ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Any = dct.pop(UpperCamelCase__ ) _a : Dict = val def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : Tuple = [] for k in state_dict.keys(): _a : Any = k if ".pwconv" in k: _a : int = k_new.replace(""".pwconv""" , """.point_wise_conv""" ) if ".dwconv" in k: _a : List[str] = k_new.replace(""".dwconv""" , """.depth_wise_conv""" ) if ".Proj." in k: _a : Optional[int] = k_new.replace(""".Proj.""" , """.proj.""" ) if "patch_embed" in k_new: _a : Tuple = k_new.replace("""patch_embed""" , """swiftformer.patch_embed.patch_embedding""" ) if "network" in k_new: _a : int = k_new.split(""".""" ) if ls[2].isdigit(): _a : Union[str, Any] = """swiftformer.encoder.network.""" + ls[1] + """.blocks.""" + ls[2] + """.""" + """.""".join(ls[3:] ) else: _a : Tuple = k_new.replace("""network""" , """swiftformer.encoder.network""" ) rename_keys.append((k, k_new) ) return rename_keys @torch.no_grad() def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Tuple = SwiftFormerConfig() # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size _a : Optional[int] = 1_0_0_0 _a : Optional[Any] = """huggingface/label-files""" _a : Optional[Any] = """imagenet-1k-id2label.json""" _a : List[str] = json.load(open(hf_hub_download(UpperCamelCase__ , UpperCamelCase__ , repo_type="""dataset""" ) , """r""" ) ) _a : Optional[Any] = {int(UpperCamelCase__ ): v for k, v in idalabel.items()} _a : Dict = idalabel _a : Optional[int] = {v: k for k, v in idalabel.items()} # size of the architecture if swiftformer_name == "swiftformer_xs": _a : Any = [3, 3, 6, 4] _a : int = [4_8, 5_6, 1_1_2, 2_2_0] elif swiftformer_name == "swiftformer_s": _a : Any = [3, 3, 9, 6] _a : List[str] = [4_8, 6_4, 1_6_8, 2_2_4] elif swiftformer_name == "swiftformer_l1": _a : List[Any] = [4, 3, 1_0, 5] _a : Optional[int] = [4_8, 9_6, 1_9_2, 3_8_4] elif swiftformer_name == "swiftformer_l3": _a : List[Any] = [4, 4, 1_2, 6] _a : Optional[Any] = [6_4, 1_2_8, 3_2_0, 5_1_2] # load state_dict of original model, remove and rename some keys if original_ckpt: if original_ckpt.startswith("""https""" ): _a : Tuple = torch.hub.load_state_dict_from_url(UpperCamelCase__ , map_location="""cpu""" , check_hash=UpperCamelCase__ ) else: _a : Dict = torch.load(UpperCamelCase__ , map_location="""cpu""" ) _a : int = checkpoint _a : Optional[Any] = create_rename_keys(UpperCamelCase__ ) for rename_key_src, rename_key_dest in rename_keys: rename_key(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # load HuggingFace model _a : Any = SwiftFormerForImageClassification(UpperCamelCase__ ).eval() hf_model.load_state_dict(UpperCamelCase__ ) # prepare test inputs _a : Any = prepare_img() _a : Union[str, Any] = ViTImageProcessor.from_pretrained("""preprocessor_config""" ) _a : Optional[int] = processor(images=UpperCamelCase__ , return_tensors="""pt""" ) # compare outputs from both models _a : Dict = get_expected_output(UpperCamelCase__ ) _a : int = hf_model(inputs["""pixel_values"""] ).logits assert hf_logits.shape == torch.Size([1, 1_0_0_0] ) assert torch.allclose(hf_logits[0, 0:5] , UpperCamelCase__ , atol=1e-3 ) Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) print(F"""Saving model {swiftformer_name} to {pytorch_dump_folder_path}""" ) hf_model.save_pretrained(UpperCamelCase__ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '--swiftformer_name', default='swiftformer_xs', choices=['swiftformer_xs', 'swiftformer_s', 'swiftformer_l1', 'swiftformer_l3'], type=str, help='Name of the SwiftFormer model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default='./converted_outputs/', type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument('--original_ckpt', default=None, type=str, help='Path to the original model checkpoint.') _snake_case = parser.parse_args() convert_swiftformer_checkpoint(args.swiftformer_name, args.pytorch_dump_folder_path, args.original_ckpt)
294
1
"""simple docstring""" from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer _snake_case = logging.get_logger(__name__) _snake_case = { 'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_config_file': 'tokenizer_config.json', } _snake_case = { 'vocab_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json' }, 'merges_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt' }, 'tokenizer_config_file': { 'facebook/blenderbot_small-90M': ( 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json' ) }, } _snake_case = { 'facebook/blenderbot_small-90M': 512, } class UpperCamelCase ( snake_case_ ): UpperCamelCase : List[Any] = VOCAB_FILES_NAMES UpperCamelCase : int = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase : List[Any] = BlenderbotSmallTokenizer def __init__( self : int , UpperCAmelCase__ : Any=None , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : Any="<|endoftext|>" , UpperCAmelCase__ : List[str]="<|endoftext|>" , UpperCAmelCase__ : Any="<|endoftext|>" , UpperCAmelCase__ : Dict=False , UpperCAmelCase__ : List[Any]=True , **UpperCAmelCase__ : int , ) -> str: super().__init__( ByteLevelBPETokenizer( vocab=UpperCAmelCase__ , merges=UpperCAmelCase__ , add_prefix_space=UpperCAmelCase__ , trim_offsets=UpperCAmelCase__ , ) , bos_token=UpperCAmelCase__ , eos_token=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , **UpperCAmelCase__ , ) _a : Dict = add_prefix_space def _lowercase ( self : List[Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : List[str]=None ) -> int: _a : Optional[Any] = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def _lowercase ( self : Dict , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ) -> List[int]: _a : Tuple = [self.sep_token_id] _a : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
294
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case = { 'configuration_perceiver': ['PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PerceiverConfig', 'PerceiverOnnxConfig'], 'tokenization_perceiver': ['PerceiverTokenizer'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ['PerceiverFeatureExtractor'] _snake_case = ['PerceiverImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST', 'PerceiverForImageClassificationConvProcessing', 'PerceiverForImageClassificationFourier', 'PerceiverForImageClassificationLearned', 'PerceiverForMaskedLM', 'PerceiverForMultimodalAutoencoding', 'PerceiverForOpticalFlow', 'PerceiverForSequenceClassification', 'PerceiverLayer', 'PerceiverModel', 'PerceiverPreTrainedModel', ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
294
1
"""simple docstring""" import os import unittest from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, BertTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class UpperCamelCase ( snake_case_ , unittest.TestCase ): UpperCamelCase : Optional[int] = BertTokenizer UpperCamelCase : str = BertTokenizerFast UpperCamelCase : List[Any] = True UpperCamelCase : str = True UpperCamelCase : Any = filter_non_english def _lowercase ( self : int ) -> List[Any]: super().setUp() _a : Union[str, Any] = [ """[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest""", ] _a : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def _lowercase ( self : Tuple , UpperCAmelCase__ : str ) -> Any: _a : Dict = """UNwant\u00E9d,running""" _a : Tuple = """unwanted, running""" return input_text, output_text def _lowercase ( self : Optional[int] ) -> Optional[Any]: _a : Optional[Any] = self.tokenizer_class(self.vocab_file ) _a : Dict = tokenizer.tokenize("""UNwant\u00E9d,running""" ) self.assertListEqual(UpperCAmelCase__ , ["""un""", """##want""", """##ed""", """,""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) , [9, 6, 7, 12, 10, 11] ) def _lowercase ( self : Tuple ) -> List[Any]: if not self.test_rust_tokenizer: return _a : Union[str, Any] = self.get_tokenizer() _a : Tuple = self.get_rust_tokenizer() _a : str = """UNwant\u00E9d,running""" _a : Union[str, Any] = tokenizer.tokenize(UpperCAmelCase__ ) _a : Any = rust_tokenizer.tokenize(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[Any] = tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) _a : Dict = rust_tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) _a : int = self.get_rust_tokenizer() _a : int = tokenizer.encode(UpperCAmelCase__ ) _a : List[Any] = rust_tokenizer.encode(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) # With lower casing _a : str = self.get_tokenizer(do_lower_case=UpperCAmelCase__ ) _a : Tuple = self.get_rust_tokenizer(do_lower_case=UpperCAmelCase__ ) _a : List[str] = """UNwant\u00E9d,running""" _a : Optional[Any] = tokenizer.tokenize(UpperCAmelCase__ ) _a : List[Any] = rust_tokenizer.tokenize(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[str] = tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) _a : Union[str, Any] = rust_tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Optional[int] = self.get_rust_tokenizer() _a : Any = tokenizer.encode(UpperCAmelCase__ ) _a : List[str] = rust_tokenizer.encode(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def _lowercase ( self : Union[str, Any] ) -> List[str]: _a : str = BasicTokenizer() self.assertListEqual(tokenizer.tokenize("""ah\u535A\u63A8zz""" ) , ["""ah""", """\u535A""", """\u63A8""", """zz"""] ) def _lowercase ( self : List[Any] ) -> Optional[int]: _a : Optional[Any] = BasicTokenizer(do_lower_case=UpperCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""hello""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def _lowercase ( self : int ) -> Tuple: _a : Optional[Any] = BasicTokenizer(do_lower_case=UpperCAmelCase__ , strip_accents=UpperCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hällo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""h\u00E9llo"""] ) def _lowercase ( self : Any ) -> Dict: _a : str = BasicTokenizer(do_lower_case=UpperCAmelCase__ , strip_accents=UpperCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def _lowercase ( self : str ) -> Dict: _a : Optional[int] = BasicTokenizer(do_lower_case=UpperCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def _lowercase ( self : str ) -> List[str]: _a : Union[str, Any] = BasicTokenizer(do_lower_case=UpperCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _lowercase ( self : Optional[Any] ) -> int: _a : Optional[int] = BasicTokenizer(do_lower_case=UpperCAmelCase__ , strip_accents=UpperCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HäLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _lowercase ( self : Optional[int] ) -> Union[str, Any]: _a : str = BasicTokenizer(do_lower_case=UpperCAmelCase__ , strip_accents=UpperCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HaLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _lowercase ( self : Optional[int] ) -> Tuple: _a : Tuple = BasicTokenizer(do_lower_case=UpperCAmelCase__ , never_split=["""[UNK]"""] ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? [UNK]""" ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?""", """[UNK]"""] ) def _lowercase ( self : Optional[Any] ) -> int: _a : str = BasicTokenizer() _a : Union[str, Any] = """a\n'll !!to?'d of, can't.""" _a : Optional[Any] = ["""a""", """'""", """ll""", """!""", """!""", """to""", """?""", """'""", """d""", """of""", """,""", """can""", """'""", """t""", """."""] self.assertListEqual(tokenizer.tokenize(UpperCAmelCase__ ) , UpperCAmelCase__ ) def _lowercase ( self : List[str] ) -> Dict: _a : List[str] = ["""[UNK]""", """[CLS]""", """[SEP]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing"""] _a : Tuple = {} for i, token in enumerate(UpperCAmelCase__ ): _a : Optional[Any] = i _a : List[Any] = WordpieceTokenizer(vocab=UpperCAmelCase__ , unk_token="""[UNK]""" ) self.assertListEqual(tokenizer.tokenize("""""" ) , [] ) self.assertListEqual(tokenizer.tokenize("""unwanted running""" ) , ["""un""", """##want""", """##ed""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.tokenize("""unwantedX running""" ) , ["""[UNK]""", """runn""", """##ing"""] ) def _lowercase ( self : int ) -> List[str]: self.assertTrue(_is_whitespace(""" """ ) ) self.assertTrue(_is_whitespace("""\t""" ) ) self.assertTrue(_is_whitespace("""\r""" ) ) self.assertTrue(_is_whitespace("""\n""" ) ) self.assertTrue(_is_whitespace("""\u00A0""" ) ) self.assertFalse(_is_whitespace("""A""" ) ) self.assertFalse(_is_whitespace("""-""" ) ) def _lowercase ( self : Dict ) -> str: self.assertTrue(_is_control("""\u0005""" ) ) self.assertFalse(_is_control("""A""" ) ) self.assertFalse(_is_control(""" """ ) ) self.assertFalse(_is_control("""\t""" ) ) self.assertFalse(_is_control("""\r""" ) ) def _lowercase ( self : Any ) -> str: self.assertTrue(_is_punctuation("""-""" ) ) self.assertTrue(_is_punctuation("""$""" ) ) self.assertTrue(_is_punctuation("""`""" ) ) self.assertTrue(_is_punctuation(""".""" ) ) self.assertFalse(_is_punctuation("""A""" ) ) self.assertFalse(_is_punctuation(""" """ ) ) def _lowercase ( self : Union[str, Any] ) -> Tuple: _a : Optional[int] = self.get_tokenizer() _a : Optional[Any] = self.get_rust_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(UpperCAmelCase__ ) for t in ["""Test""", """\xad""", """test"""]] , [["""[UNK]"""], [], ["""[UNK]"""]] ) self.assertListEqual( [rust_tokenizer.tokenize(UpperCAmelCase__ ) for t in ["""Test""", """\xad""", """test"""]] , [["""[UNK]"""], [], ["""[UNK]"""]] ) @slow def _lowercase ( self : Tuple ) -> Optional[int]: _a : Any = self.tokenizer_class.from_pretrained("""bert-base-uncased""" ) _a : Any = tokenizer.encode("""sequence builders""" , add_special_tokens=UpperCAmelCase__ ) _a : Optional[int] = tokenizer.encode("""multi-sequence build""" , add_special_tokens=UpperCAmelCase__ ) _a : Tuple = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase__ ) _a : Union[str, Any] = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase__ , UpperCAmelCase__ ) assert encoded_sentence == [101] + text + [102] assert encoded_pair == [101] + text + [102] + text_a + [102] def _lowercase ( self : Dict ) -> int: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _a : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__ ) _a : Optional[Any] = f"""A, naïve {tokenizer_r.mask_token} AllenNLP sentence.""" _a : str = tokenizer_r.encode_plus( UpperCAmelCase__ , return_attention_mask=UpperCAmelCase__ , return_token_type_ids=UpperCAmelCase__ , return_offsets_mapping=UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ , ) _a : List[str] = tokenizer_r.do_lower_case if hasattr(UpperCAmelCase__ , """do_lower_case""" ) else False _a : List[str] = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """A"""), ((1, 2), ""","""), ((3, 5), """na"""), ((5, 6), """##ï"""), ((6, 8), """##ve"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """Allen"""), ((21, 23), """##NL"""), ((23, 24), """##P"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """a"""), ((1, 2), ""","""), ((3, 8), """naive"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """allen"""), ((21, 23), """##nl"""), ((23, 24), """##p"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens["""input_ids"""] ) ) self.assertEqual([e[0] for e in expected_results] , tokens["""offset_mapping"""] ) def _lowercase ( self : Optional[int] ) -> Any: _a : str = ["""的""", """人""", """有"""] _a : Optional[int] = """""".join(UpperCAmelCase__ ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _a : List[str] = True _a : Union[str, Any] = self.tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__ ) _a : Optional[Any] = self.rust_tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__ ) _a : List[str] = tokenizer_p.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) _a : str = tokenizer_r.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) _a : int = tokenizer_r.convert_ids_to_tokens(UpperCAmelCase__ ) _a : Optional[int] = tokenizer_p.convert_ids_to_tokens(UpperCAmelCase__ ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[str] = False _a : List[Any] = self.rust_tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__ ) _a : Union[str, Any] = self.tokenizer_class.from_pretrained(UpperCAmelCase__ , **UpperCAmelCase__ ) _a : Any = tokenizer_r.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) _a : int = tokenizer_p.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) _a : Optional[int] = tokenizer_r.convert_ids_to_tokens(UpperCAmelCase__ ) _a : List[Any] = tokenizer_p.convert_ids_to_tokens(UpperCAmelCase__ ) # it is expected that only the first Chinese character is not preceded by "##". _a : Any = [ f"""##{token}""" if idx != 0 else token for idx, token in enumerate(UpperCAmelCase__ ) ] self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ )
294
"""simple docstring""" import cva import numpy as np class UpperCamelCase : def __init__( self : Optional[int] , UpperCAmelCase__ : float , UpperCAmelCase__ : int ) -> Dict: if k in (0.0_4, 0.0_6): _a : List[str] = k _a : List[Any] = window_size else: raise ValueError("""invalid k value""" ) def __str__( self : Dict ) -> str: return str(self.k ) def _lowercase ( self : int , UpperCAmelCase__ : str ) -> tuple[cva.Mat, list[list[int]]]: _a : Dict = cva.imread(UpperCAmelCase__ , 0 ) _a , _a : List[Any] = img.shape _a : list[list[int]] = [] _a : List[Any] = img.copy() _a : int = cva.cvtColor(UpperCAmelCase__ , cva.COLOR_GRAY2RGB ) _a , _a : Any = np.gradient(UpperCAmelCase__ ) _a : Tuple = dx**2 _a : Union[str, Any] = dy**2 _a : Union[str, Any] = dx * dy _a : int = 0.0_4 _a : List[str] = self.window_size // 2 for y in range(UpperCAmelCase__ , h - offset ): for x in range(UpperCAmelCase__ , w - offset ): _a : str = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _a : List[Any] = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _a : Tuple = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _a : Any = (wxx * wyy) - (wxy**2) _a : Tuple = wxx + wyy _a : Any = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 255 ) return color_img, corner_list if __name__ == "__main__": _snake_case = HarrisCorner(0.04, 3) _snake_case , _snake_case = edge_detect.detect('path_to_image') cva.imwrite('detect.png', color_img)
294
1
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available, is_torch_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, is_pt_tf_cross_test, slow if is_tf_available(): from transformers import ( AutoConfig, BertConfig, GPTaConfig, TaConfig, TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST if is_torch_available(): from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelWithLMHead, BertForMaskedLM, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertModel, GPTaLMHeadModel, RobertaForMaskedLM, TaForConditionalGeneration, ) @is_pt_tf_cross_test class UpperCamelCase ( unittest.TestCase ): @slow def _lowercase ( self : str ) -> Any: # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: _a : Optional[Any] = AutoConfig.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Optional[int] = TFAutoModel.from_pretrained(UpperCAmelCase__ , from_pt=UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Any = AutoModel.from_pretrained(UpperCAmelCase__ , from_tf=UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) @slow def _lowercase ( self : Optional[int] ) -> Dict: # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: _a : Optional[int] = AutoConfig.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[str] = TFAutoModelForPreTraining.from_pretrained(UpperCAmelCase__ , from_pt=UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[str] = AutoModelForPreTraining.from_pretrained(UpperCAmelCase__ , from_tf=UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) @slow def _lowercase ( self : int ) -> List[str]: for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _a : Union[str, Any] = AutoConfig.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Union[str, Any] = TFAutoModelForCausalLM.from_pretrained(UpperCAmelCase__ , from_pt=UpperCAmelCase__ ) _a , _a : Any = TFAutoModelForCausalLM.from_pretrained( UpperCAmelCase__ , output_loading_info=UpperCAmelCase__ , from_pt=UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[str] = AutoModelForCausalLM.from_pretrained(UpperCAmelCase__ , from_tf=UpperCAmelCase__ ) _a , _a : Any = AutoModelForCausalLM.from_pretrained( UpperCAmelCase__ , output_loading_info=UpperCAmelCase__ , from_tf=UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) @slow def _lowercase ( self : Any ) -> List[str]: for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _a : Tuple = AutoConfig.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Tuple = TFAutoModelWithLMHead.from_pretrained(UpperCAmelCase__ , from_pt=UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) _a : str = AutoModelWithLMHead.from_pretrained(UpperCAmelCase__ , from_tf=UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) @slow def _lowercase ( self : int ) -> Any: for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _a : str = AutoConfig.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[Any] = TFAutoModelForMaskedLM.from_pretrained(UpperCAmelCase__ , from_pt=UpperCAmelCase__ ) _a , _a : Any = TFAutoModelForMaskedLM.from_pretrained( UpperCAmelCase__ , output_loading_info=UpperCAmelCase__ , from_pt=UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[Any] = AutoModelForMaskedLM.from_pretrained(UpperCAmelCase__ , from_tf=UpperCAmelCase__ ) _a , _a : Any = AutoModelForMaskedLM.from_pretrained( UpperCAmelCase__ , output_loading_info=UpperCAmelCase__ , from_tf=UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) @slow def _lowercase ( self : Optional[int] ) -> str: for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _a : Any = AutoConfig.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Tuple = TFAutoModelForSeqaSeqLM.from_pretrained(UpperCAmelCase__ , from_pt=UpperCAmelCase__ ) _a , _a : Optional[Any] = TFAutoModelForSeqaSeqLM.from_pretrained( UpperCAmelCase__ , output_loading_info=UpperCAmelCase__ , from_pt=UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) _a : str = AutoModelForSeqaSeqLM.from_pretrained(UpperCAmelCase__ , from_tf=UpperCAmelCase__ ) _a , _a : int = AutoModelForSeqaSeqLM.from_pretrained( UpperCAmelCase__ , output_loading_info=UpperCAmelCase__ , from_tf=UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) @slow def _lowercase ( self : int ) -> Tuple: # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: _a : str = AutoConfig.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Optional[Any] = TFAutoModelForSequenceClassification.from_pretrained(UpperCAmelCase__ , from_pt=UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[Any] = AutoModelForSequenceClassification.from_pretrained(UpperCAmelCase__ , from_tf=UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) @slow def _lowercase ( self : Union[str, Any] ) -> Dict: # for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["bert-base-uncased"]: _a : Optional[int] = AutoConfig.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Union[str, Any] = TFAutoModelForQuestionAnswering.from_pretrained(UpperCAmelCase__ , from_pt=UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[Any] = AutoModelForQuestionAnswering.from_pretrained(UpperCAmelCase__ , from_tf=UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) def _lowercase ( self : int ) -> Optional[int]: _a : Optional[int] = TFAutoModelWithLMHead.from_pretrained(UpperCAmelCase__ , from_pt=UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=UpperCAmelCase__ ) , 14410 ) _a : List[Any] = AutoModelWithLMHead.from_pretrained(UpperCAmelCase__ , from_tf=UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=UpperCAmelCase__ ) , 14410 ) def _lowercase ( self : List[Any] ) -> Any: _a : Tuple = TFAutoModelWithLMHead.from_pretrained(UpperCAmelCase__ , from_pt=UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=UpperCAmelCase__ ) , 14410 ) _a : str = AutoModelWithLMHead.from_pretrained(UpperCAmelCase__ , from_tf=UpperCAmelCase__ ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=UpperCAmelCase__ ) , 14410 )
294
"""simple docstring""" def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' # Check if the input is valid if not len(UpperCamelCase__ ) == len(UpperCamelCase__ ) == 3: raise ValueError("""Please enter a valid equation.""" ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError("""Both a & b of two equations can't be zero.""" ) # Extract the coefficients _a , _a , _a : Any = equationa _a , _a , _a : Tuple = equationa # Calculate the determinants of the matrices _a : int = aa * ba - aa * ba _a : str = ca * ba - ca * ba _a : str = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError("""Infinite solutions. (Consistent system)""" ) else: raise ValueError("""No solution. (Inconsistent system)""" ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: _a : Dict = determinant_x / determinant _a : str = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
294
1
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ShapEPipeline else: from .camera import create_pan_cameras from .pipeline_shap_e import ShapEPipeline from .pipeline_shap_e_img2img import ShapEImgaImgPipeline from .renderer import ( BoundingBoxVolume, ImportanceRaySampler, MLPNeRFModelOutput, MLPNeRSTFModel, ShapEParamsProjModel, ShapERenderer, StratifiedRaySampler, VoidNeRFModel, )
294
"""simple docstring""" _snake_case = '\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n' _snake_case = [{'type': 'code', 'content': INSTALL_CONTENT}] _snake_case = { '{processor_class}': 'FakeProcessorClass', '{model_class}': 'FakeModelClass', '{object_class}': 'FakeObjectClass', }
294
1
"""simple docstring""" import copy import inspect import unittest from transformers import PretrainedConfig, SwiftFormerConfig from transformers.testing_utils import ( require_torch, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwiftFormerForImageClassification, SwiftFormerModel from transformers.models.swiftformer.modeling_swiftformer import SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class UpperCamelCase : def __init__( self : Optional[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Union[str, Any]=13 , UpperCAmelCase__ : int=3 , UpperCAmelCase__ : List[Any]=True , UpperCAmelCase__ : Optional[Any]=True , UpperCAmelCase__ : Any=0.1 , UpperCAmelCase__ : Union[str, Any]=0.1 , UpperCAmelCase__ : Union[str, Any]=224 , UpperCAmelCase__ : Optional[int]=1000 , UpperCAmelCase__ : Tuple=[3, 3, 6, 4] , UpperCAmelCase__ : str=[48, 56, 112, 220] , ) -> int: _a : Optional[Any] = parent _a : Dict = batch_size _a : Union[str, Any] = num_channels _a : Any = is_training _a : Optional[Any] = use_labels _a : Union[str, Any] = hidden_dropout_prob _a : Dict = attention_probs_dropout_prob _a : Any = num_labels _a : Optional[Any] = image_size _a : str = layer_depths _a : int = embed_dims def _lowercase ( self : int ) -> Optional[Any]: _a : Dict = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _a : Dict = None if self.use_labels: _a : Union[str, Any] = ids_tensor([self.batch_size] , self.num_labels ) _a : Any = self.get_config() return config, pixel_values, labels def _lowercase ( self : Optional[int] ) -> List[str]: return SwiftFormerConfig( depths=self.layer_depths , embed_dims=self.embed_dims , mlp_ratio=4 , downsamples=[True, True, True, True] , hidden_act="""gelu""" , num_labels=self.num_labels , down_patch_size=3 , down_stride=2 , down_pad=1 , drop_rate=0.0 , drop_path_rate=0.0 , use_layer_scale=UpperCAmelCase__ , layer_scale_init_value=1E-5 , ) def _lowercase ( self : str , UpperCAmelCase__ : int , UpperCAmelCase__ : str , UpperCAmelCase__ : int ) -> List[Any]: _a : List[str] = SwiftFormerModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() _a : Tuple = model(UpperCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dims[-1], 7, 7) ) def _lowercase ( self : Union[str, Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Any ) -> List[str]: _a : Tuple = self.num_labels _a : Optional[Any] = SwiftFormerForImageClassification(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() _a : Union[str, Any] = model(UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) _a : List[str] = SwiftFormerForImageClassification(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() _a : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _a : List[str] = model(UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowercase ( self : List[Any] ) -> Dict: ((_a) , (_a) , (_a)) : Any = self.prepare_config_and_inputs() _a : str = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class UpperCamelCase ( snake_case_ , snake_case_ , unittest.TestCase ): UpperCamelCase : Any = (SwiftFormerModel, SwiftFormerForImageClassification) if is_torch_available() else () UpperCamelCase : Dict = ( {'''feature-extraction''': SwiftFormerModel, '''image-classification''': SwiftFormerForImageClassification} if is_torch_available() else {} ) UpperCamelCase : Union[str, Any] = False UpperCamelCase : Dict = False UpperCamelCase : Dict = False UpperCamelCase : str = False UpperCamelCase : Union[str, Any] = False def _lowercase ( self : str ) -> Union[str, Any]: _a : int = SwiftFormerModelTester(self ) _a : int = ConfigTester( self , config_class=UpperCAmelCase__ , has_text_modality=UpperCAmelCase__ , hidden_size=37 , num_attention_heads=12 , num_hidden_layers=12 , ) def _lowercase ( self : List[Any] ) -> Optional[int]: self.config_tester.run_common_tests() @unittest.skip(reason="""SwiftFormer does not use inputs_embeds""" ) def _lowercase ( self : Union[str, Any] ) -> List[str]: pass def _lowercase ( self : Optional[Any] ) -> List[str]: _a , _a : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Optional[Any] = model_class(UpperCAmelCase__ ) _a : Tuple = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCAmelCase__ , nn.Linear ) ) def _lowercase ( self : List[Any] ) -> Dict: _a , _a : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Optional[Any] = model_class(UpperCAmelCase__ ) _a : Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _a : Any = [*signature.parameters.keys()] _a : List[str] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCAmelCase__ ) def _lowercase ( self : List[str] ) -> str: _a : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase__ ) def _lowercase ( self : List[Any] ) -> str: _a : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCAmelCase__ ) @slow def _lowercase ( self : Dict ) -> Any: for model_name in SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _a : List[Any] = SwiftFormerModel.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) @unittest.skip(reason="""SwiftFormer does not output attentions""" ) def _lowercase ( self : List[str] ) -> List[str]: pass def _lowercase ( self : str ) -> int: def check_hidden_states_output(UpperCAmelCase__ : List[str] , UpperCAmelCase__ : int , UpperCAmelCase__ : Any ): _a : Tuple = model_class(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() with torch.no_grad(): _a : List[Any] = model(**self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ ) ) _a : List[Any] = outputs.hidden_states _a : Optional[int] = 8 self.assertEqual(len(UpperCAmelCase__ ) , UpperCAmelCase__ ) # TODO # SwiftFormer's feature maps are of shape (batch_size, embed_dims, height, width) # with the width and height being successively divided by 2, after every 2 blocks for i in range(len(UpperCAmelCase__ ) ): self.assertEqual( hidden_states[i].shape , torch.Size( [ self.model_tester.batch_size, self.model_tester.embed_dims[i // 2], (self.model_tester.image_size // 4) // 2 ** (i // 2), (self.model_tester.image_size // 4) // 2 ** (i // 2), ] ) , ) _a , _a : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : str = True check_hidden_states_output(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _a : Tuple = True check_hidden_states_output(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) def _lowercase ( self : Optional[Any] ) -> Union[str, Any]: def _config_zero_init(UpperCAmelCase__ : List[str] ): _a : Union[str, Any] = copy.deepcopy(UpperCAmelCase__ ) for key in configs_no_init.__dict__.keys(): if "_range" in key or "_std" in key or "initializer_factor" in key or "layer_scale" in key: setattr(UpperCAmelCase__ , UpperCAmelCase__ , 1E-10 ) if isinstance(getattr(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) , UpperCAmelCase__ ): _a : Tuple = _config_zero_init(getattr(UpperCAmelCase__ , UpperCAmelCase__ ) ) setattr(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) return configs_no_init _a , _a : Any = self.model_tester.prepare_config_and_inputs_for_common() _a : Any = _config_zero_init(UpperCAmelCase__ ) for model_class in self.all_model_classes: _a : str = model_class(config=UpperCAmelCase__ ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1E9) / 1E9).round().item() , [0.0, 1.0] , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def _lowercase ( self : Optional[int] ) -> Union[str, Any]: pass def lowerCAmelCase__ ( ): '''simple docstring''' _a : Optional[int] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class UpperCamelCase ( unittest.TestCase ): @cached_property def _lowercase ( self : Optional[Any] ) -> List[Any]: return ViTImageProcessor.from_pretrained("""MBZUAI/swiftformer-xs""" ) if is_vision_available() else None @slow def _lowercase ( self : int ) -> Optional[int]: _a : Dict = SwiftFormerForImageClassification.from_pretrained("""MBZUAI/swiftformer-xs""" ).to(UpperCAmelCase__ ) _a : int = self.default_image_processor _a : Optional[int] = prepare_img() _a : Tuple = image_processor(images=UpperCAmelCase__ , return_tensors="""pt""" ).to(UpperCAmelCase__ ) # forward pass with torch.no_grad(): _a : Any = model(**UpperCAmelCase__ ) # verify the logits _a : str = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , UpperCAmelCase__ ) _a : Optional[Any] = torch.tensor([[-2.1703E00, 2.1107E00, -2.0811E00]] ).to(UpperCAmelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCAmelCase__ , atol=1E-4 ) )
294
"""simple docstring""" import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json', } class UpperCamelCase ( snake_case_ ): UpperCamelCase : int = '''mvp''' UpperCamelCase : Union[str, Any] = ['''past_key_values'''] UpperCamelCase : Any = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : List[str] , UpperCAmelCase__ : List[str]=50267 , UpperCAmelCase__ : Optional[Any]=1024 , UpperCAmelCase__ : Tuple=12 , UpperCAmelCase__ : Optional[Any]=4096 , UpperCAmelCase__ : int=16 , UpperCAmelCase__ : Tuple=12 , UpperCAmelCase__ : int=4096 , UpperCAmelCase__ : List[Any]=16 , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : Tuple="gelu" , UpperCAmelCase__ : Union[str, Any]=1024 , UpperCAmelCase__ : List[str]=0.1 , UpperCAmelCase__ : Any=0.0 , UpperCAmelCase__ : Dict=0.0 , UpperCAmelCase__ : Tuple=0.0_2 , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : Optional[Any]=False , UpperCAmelCase__ : int=True , UpperCAmelCase__ : Tuple=1 , UpperCAmelCase__ : Dict=0 , UpperCAmelCase__ : Union[str, Any]=2 , UpperCAmelCase__ : Optional[int]=True , UpperCAmelCase__ : Tuple=2 , UpperCAmelCase__ : Any=2 , UpperCAmelCase__ : Optional[Any]=False , UpperCAmelCase__ : Dict=100 , UpperCAmelCase__ : Union[str, Any]=800 , **UpperCAmelCase__ : Dict , ) -> List[Any]: _a : Any = vocab_size _a : Any = max_position_embeddings _a : Union[str, Any] = d_model _a : List[str] = encoder_ffn_dim _a : List[Any] = encoder_layers _a : Dict = encoder_attention_heads _a : Tuple = decoder_ffn_dim _a : List[Any] = decoder_layers _a : Optional[Any] = decoder_attention_heads _a : Optional[Any] = dropout _a : str = attention_dropout _a : Dict = activation_dropout _a : Any = activation_function _a : Tuple = init_std _a : Dict = encoder_layerdrop _a : Optional[int] = decoder_layerdrop _a : Optional[Any] = classifier_dropout _a : List[Any] = use_cache _a : Dict = encoder_layers _a : str = scale_embedding # scale factor will be sqrt(d_model) if True _a : int = use_prompt _a : Dict = prompt_length _a : Dict = prompt_mid_dim super().__init__( pad_token_id=UpperCAmelCase__ , bos_token_id=UpperCAmelCase__ , eos_token_id=UpperCAmelCase__ , is_encoder_decoder=UpperCAmelCase__ , decoder_start_token_id=UpperCAmelCase__ , forced_eos_token_id=UpperCAmelCase__ , **UpperCAmelCase__ , ) if self.forced_bos_token_id is None and kwargs.get("""force_bos_token_to_be_generated""" , UpperCAmelCase__ ): _a : List[str] = self.bos_token_id warnings.warn( f"""Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. """ """The config can simply be saved and uploaded again to be fixed.""" )
294
1
"""simple docstring""" from typing import Optional from urllib.parse import quote import huggingface_hub as hfh from packaging import version def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = None ): '''simple docstring''' if version.parse(hfh.__version__ ).release < version.parse("""0.11.0""" ).release: # old versions of hfh don't url-encode the file path _a : Dict = quote(UpperCamelCase__ ) return hfh.hf_hub_url(UpperCamelCase__ , UpperCamelCase__ , repo_type="""dataset""" , revision=UpperCamelCase__ )
294
"""simple docstring""" import logging import os from typing import List, TextIO, Union from conllu import parse_incr from utils_ner import InputExample, Split, TokenClassificationTask _snake_case = logging.getLogger(__name__) class UpperCamelCase ( snake_case_ ): def __init__( self : Optional[Any] , UpperCAmelCase__ : Optional[int]=-1 ) -> Tuple: # in NER datasets, the last column is usually reserved for NER label _a : Optional[int] = label_idx def _lowercase ( self : Any , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Union[Split, str] ) -> List[InputExample]: if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): _a : Any = mode.value _a : Optional[int] = os.path.join(UpperCAmelCase__ , f"""{mode}.txt""" ) _a : int = 1 _a : int = [] with open(UpperCAmelCase__ , encoding="""utf-8""" ) as f: _a : str = [] _a : str = [] for line in f: if line.startswith("""-DOCSTART-""" ) or line == "" or line == "\n": if words: examples.append(InputExample(guid=f"""{mode}-{guid_index}""" , words=UpperCAmelCase__ , labels=UpperCAmelCase__ ) ) guid_index += 1 _a : List[str] = [] _a : str = [] else: _a : List[Any] = line.split(""" """ ) words.append(splits[0] ) if len(UpperCAmelCase__ ) > 1: labels.append(splits[self.label_idx].replace("""\n""" , """""" ) ) else: # Examples could have no label for mode = "test" labels.append("""O""" ) if words: examples.append(InputExample(guid=f"""{mode}-{guid_index}""" , words=UpperCAmelCase__ , labels=UpperCAmelCase__ ) ) return examples def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : List ) -> Union[str, Any]: _a : List[str] = 0 for line in test_input_reader: if line.startswith("""-DOCSTART-""" ) or line == "" or line == "\n": writer.write(UpperCAmelCase__ ) if not preds_list[example_id]: example_id += 1 elif preds_list[example_id]: _a : int = line.split()[0] + """ """ + preds_list[example_id].pop(0 ) + """\n""" writer.write(UpperCAmelCase__ ) else: logger.warning("""Maximum sequence length exceeded: No prediction for '%s'.""" , line.split()[0] ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : str ) -> List[str]: if path: with open(UpperCAmelCase__ , """r""" ) as f: _a : List[Any] = f.read().splitlines() if "O" not in labels: _a : Union[str, Any] = ["""O"""] + labels return labels else: return ["O", "B-MISC", "I-MISC", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC"] class UpperCamelCase ( snake_case_ ): def __init__( self : Union[str, Any] ) -> List[str]: # in CONLL2003 dataset chunk column is second-to-last super().__init__(label_idx=-2 ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : str ) -> List[str]: if path: with open(UpperCAmelCase__ , """r""" ) as f: _a : Optional[int] = f.read().splitlines() if "O" not in labels: _a : Optional[Any] = ["""O"""] + labels return labels else: return [ "O", "B-ADVP", "B-INTJ", "B-LST", "B-PRT", "B-NP", "B-SBAR", "B-VP", "B-ADJP", "B-CONJP", "B-PP", "I-ADVP", "I-INTJ", "I-LST", "I-PRT", "I-NP", "I-SBAR", "I-VP", "I-ADJP", "I-CONJP", "I-PP", ] class UpperCamelCase ( snake_case_ ): def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Union[Split, str] ) -> List[InputExample]: if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): _a : List[Any] = mode.value _a : Union[str, Any] = os.path.join(UpperCAmelCase__ , f"""{mode}.txt""" ) _a : List[str] = 1 _a : Optional[Any] = [] with open(UpperCAmelCase__ , encoding="""utf-8""" ) as f: for sentence in parse_incr(UpperCAmelCase__ ): _a : List[Any] = [] _a : Any = [] for token in sentence: words.append(token["""form"""] ) labels.append(token["""upos"""] ) assert len(UpperCAmelCase__ ) == len(UpperCAmelCase__ ) if words: examples.append(InputExample(guid=f"""{mode}-{guid_index}""" , words=UpperCAmelCase__ , labels=UpperCAmelCase__ ) ) guid_index += 1 return examples def _lowercase ( self : Tuple , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : List ) -> Dict: _a : Optional[Any] = 0 for sentence in parse_incr(UpperCAmelCase__ ): _a : List[str] = preds_list[example_id] _a : str = """""" for token in sentence: out += f"""{token['form']} ({token['upos']}|{s_p.pop(0 )}) """ out += "\n" writer.write(UpperCAmelCase__ ) example_id += 1 def _lowercase ( self : List[str] , UpperCAmelCase__ : str ) -> List[str]: if path: with open(UpperCAmelCase__ , """r""" ) as f: return f.read().splitlines() else: return [ "ADJ", "ADP", "ADV", "AUX", "CCONJ", "DET", "INTJ", "NOUN", "NUM", "PART", "PRON", "PROPN", "PUNCT", "SCONJ", "SYM", "VERB", "X", ]
294
1
"""simple docstring""" def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' if len(UpperCamelCase__ ) <= 1: return [tuple(UpperCamelCase__ )] _a : str = [] def generate(UpperCamelCase__ , UpperCamelCase__ ): if k == 1: res.append(tuple(arr[:] ) ) return generate(k - 1 , UpperCamelCase__ ) for i in range(k - 1 ): if k % 2 == 0: # k is even _a , _a : int = arr[k - 1], arr[i] else: # k is odd _a , _a : Any = arr[k - 1], arr[0] generate(k - 1 , UpperCamelCase__ ) generate(len(UpperCamelCase__ ) , UpperCamelCase__ ) return res if __name__ == "__main__": _snake_case = input('Enter numbers separated by a comma:\n').strip() _snake_case = [int(item) for item in user_input.split(',')] print(heaps(arr))
294
"""simple docstring""" from __future__ import annotations import time import numpy as np _snake_case = [8, 5, 9, 7] _snake_case = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] _snake_case = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class UpperCamelCase : def __init__( self : List[Any] , UpperCAmelCase__ : list[int] , UpperCAmelCase__ : list[list[int]] , UpperCAmelCase__ : list[list[int]] , ) -> None: _a : List[str] = claim_vector _a : List[Any] = allocated_resources_table _a : Union[str, Any] = maximum_claim_table def _lowercase ( self : Tuple ) -> list[int]: return [ sum(p_item[i] for p_item in self.__allocated_resources_table ) for i in range(len(self.__allocated_resources_table[0] ) ) ] def _lowercase ( self : int ) -> list[int]: return np.array(self.__claim_vector ) - np.array( self.__processes_resource_summation() ) def _lowercase ( self : List[str] ) -> list[list[int]]: return [ list(np.array(self.__maximum_claim_table[i] ) - np.array(UpperCAmelCase__ ) ) for i, allocated_resource in enumerate(self.__allocated_resources_table ) ] def _lowercase ( self : Optional[Any] ) -> dict[int, list[int]]: return {self.__need().index(UpperCAmelCase__ ): i for i in self.__need()} def _lowercase ( self : Dict , **UpperCAmelCase__ : Optional[Any] ) -> None: _a : List[Any] = self.__need() _a : Optional[int] = self.__allocated_resources_table _a : str = self.__available_resources() _a : Optional[Any] = self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print("""_""" * 50 + """\n""" ) while need_list: _a : int = False for each_need in need_list: _a : Optional[int] = True for index, need in enumerate(UpperCAmelCase__ ): if need > available_resources[index]: _a : List[Any] = False break if execution: _a : str = True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: _a : Any = original_need_index print(f"""Process {process_number + 1} is executing.""" ) # remove the process run from stack need_list.remove(UpperCAmelCase__ ) # update available/freed resources stack _a : Union[str, Any] = np.array(UpperCAmelCase__ ) + np.array( alloc_resources_table[process_number] ) print( """Updated available resource stack for processes: """ + """ """.join([str(UpperCAmelCase__ ) for x in available_resources] ) ) break if safe: print("""The process is in a safe state.\n""" ) else: print("""System in unsafe state. Aborting...\n""" ) break def _lowercase ( self : Any ) -> Optional[int]: print(""" """ * 9 + """Allocated Resource Table""" ) for item in self.__allocated_resources_table: print( f"""P{self.__allocated_resources_table.index(UpperCAmelCase__ ) + 1}""" + """ """.join(f"""{it:>8}""" for it in item ) + """\n""" ) print(""" """ * 9 + """System Resource Table""" ) for item in self.__maximum_claim_table: print( f"""P{self.__maximum_claim_table.index(UpperCAmelCase__ ) + 1}""" + """ """.join(f"""{it:>8}""" for it in item ) + """\n""" ) print( """Current Usage by Active Processes: """ + """ """.join(str(UpperCAmelCase__ ) for x in self.__claim_vector ) ) print( """Initial Available Resources: """ + """ """.join(str(UpperCAmelCase__ ) for x in self.__available_resources() ) ) time.sleep(1 ) if __name__ == "__main__": import doctest doctest.testmod()
294
1
"""simple docstring""" import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a , _a : List[str] = image.size _a , _a : Optional[Any] = (x - x % 3_2 for x in (w, h)) # resize to integer multiple of 32 _a : str = image.resize((w, h) , resample=PIL_INTERPOLATION["""lanczos"""] ) _a : List[str] = np.array(UpperCamelCase__ ).astype(np.floataa ) / 255.0 _a : Dict = image[None].transpose(0 , 3 , 1 , 2 ) _a : List[Any] = torch.from_numpy(UpperCamelCase__ ) return 2.0 * image - 1.0 class UpperCamelCase ( snake_case_ ): def __init__( self : Optional[Any] , UpperCAmelCase__ : VQModel , UpperCAmelCase__ : UNetaDModel , UpperCAmelCase__ : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ) -> str: super().__init__() self.register_modules(vqvae=UpperCAmelCase__ , unet=UpperCAmelCase__ , scheduler=UpperCAmelCase__ ) @torch.no_grad() def __call__( self : Tuple , UpperCAmelCase__ : Union[torch.Tensor, PIL.Image.Image] = None , UpperCAmelCase__ : Optional[int] = 1 , UpperCAmelCase__ : Optional[int] = 100 , UpperCAmelCase__ : Optional[float] = 0.0 , UpperCAmelCase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , UpperCAmelCase__ : Optional[str] = "pil" , UpperCAmelCase__ : bool = True , ) -> Union[Tuple, ImagePipelineOutput]: if isinstance(UpperCAmelCase__ , PIL.Image.Image ): _a : str = 1 elif isinstance(UpperCAmelCase__ , torch.Tensor ): _a : int = image.shape[0] else: raise ValueError(f"""`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(UpperCAmelCase__ )}""" ) if isinstance(UpperCAmelCase__ , PIL.Image.Image ): _a : List[Any] = preprocess(UpperCAmelCase__ ) _a , _a : Tuple = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image _a : List[str] = (batch_size, self.unet.config.in_channels // 2, height, width) _a : Union[str, Any] = next(self.unet.parameters() ).dtype _a : List[str] = randn_tensor(UpperCAmelCase__ , generator=UpperCAmelCase__ , device=self.device , dtype=UpperCAmelCase__ ) _a : Union[str, Any] = image.to(device=self.device , dtype=UpperCAmelCase__ ) # set timesteps and move to the correct device self.scheduler.set_timesteps(UpperCAmelCase__ , device=self.device ) _a : int = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler _a : Any = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] _a : Optional[Any] = """eta""" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) _a : Optional[Any] = {} if accepts_eta: _a : Dict = eta for t in self.progress_bar(UpperCAmelCase__ ): # concat latents and low resolution image in the channel dimension. _a : Optional[Any] = torch.cat([latents, image] , dim=1 ) _a : List[Any] = self.scheduler.scale_model_input(UpperCAmelCase__ , UpperCAmelCase__ ) # predict the noise residual _a : List[str] = self.unet(UpperCAmelCase__ , UpperCAmelCase__ ).sample # compute the previous noisy sample x_t -> x_t-1 _a : int = self.scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample # decode the image latents with the VQVAE _a : int = self.vqvae.decode(UpperCAmelCase__ ).sample _a : Dict = torch.clamp(UpperCAmelCase__ , -1.0 , 1.0 ) _a : Dict = image / 2 + 0.5 _a : List[str] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": _a : Optional[int] = self.numpy_to_pil(UpperCAmelCase__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCAmelCase__ )
294
"""simple docstring""" from collections.abc import Iterable from typing import Generic, TypeVar _snake_case = TypeVar('_T') class UpperCamelCase ( Generic[_T] ): def __init__( self : Optional[int] , UpperCAmelCase__ : Iterable[_T] | None = None ) -> None: _a : list[_T] = list(iterable or [] ) _a : list[_T] = [] def __len__( self : str ) -> int: return len(self._stacka ) + len(self._stacka ) def __repr__( self : List[str] ) -> str: return f"""Queue({tuple(self._stacka[::-1] + self._stacka )})""" def _lowercase ( self : Union[str, Any] , UpperCAmelCase__ : _T ) -> None: self._stacka.append(UpperCAmelCase__ ) def _lowercase ( self : Optional[Any] ) -> _T: _a : Any = self._stacka.pop _a : Union[str, Any] = self._stacka.append if not self._stacka: while self._stacka: stacka_append(stacka_pop() ) if not self._stacka: raise IndexError("""Queue is empty""" ) return self._stacka.pop() if __name__ == "__main__": from doctest import testmod testmod()
294
1
"""simple docstring""" # Copyright (c) 2021-, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. #################################################################################################### # # Note: If when running this conversion script you're getting an exception: # ModuleNotFoundError: No module named 'megatron.model.enums' # you need to tell python where to find the clone of Megatron-LM, e.g.: # # cd /tmp # git clone https://github.com/NVIDIA/Megatron-LM # PYTHONPATH=/tmp/Megatron-LM python src/transformers/models/megatron_gpt2/convert_megatron_gpt2_checkpoint.py ... # # if you already have it cloned elsewhere, simply adjust the path to the existing path # # If the training was done using a Megatron-LM fork, e.g., # https://github.com/microsoft/Megatron-DeepSpeed/ then chances are that you need to have that one # in your path, i.e., /path/to/Megatron-DeepSpeed/ # import argparse import os import re import zipfile import torch from transformers import AutoTokenizer, GPTaConfig def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=0 ): '''simple docstring''' # Format the message. if name is None: _a : Any = None else: _a : int = """.""" * max(0 , spaces - 2 ) + """# {:""" + str(5_0 - spaces ) + """s}""" _a : Optional[int] = fmt.format(UpperCamelCase__ ) # Print and recurse (if needed). if isinstance(UpperCamelCase__ , UpperCamelCase__ ): if msg is not None: print(UpperCamelCase__ ) for k in val.keys(): recursive_print(UpperCamelCase__ , val[k] , spaces + 2 ) elif isinstance(UpperCamelCase__ , torch.Tensor ): print(UpperCamelCase__ , """:""" , val.size() ) else: print(UpperCamelCase__ , """:""" , UpperCamelCase__ ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' # Permutes layout of param tensor to [num_splits * num_heads * hidden_size, :] # for compatibility with later versions of NVIDIA Megatron-LM. # The inverse operation is performed inside Megatron-LM to read checkpoints: # https://github.com/NVIDIA/Megatron-LM/blob/v2.4/megatron/checkpointing.py#L209 # If param is the weight tensor of the self-attention block, the returned tensor # will have to be transposed one more time to be read by HuggingFace GPT2. _a : List[Any] = param.size() if checkpoint_version == 1.0: # version 1.0 stores [num_heads * hidden_size * num_splits, :] _a : Union[str, Any] = (num_heads, hidden_size, num_splits) + input_shape[1:] _a : Union[str, Any] = param.view(*UpperCamelCase__ ) _a : Optional[int] = param.transpose(0 , 2 ) _a : Union[str, Any] = param.transpose(1 , 2 ).contiguous() elif checkpoint_version >= 2.0: # other versions store [num_heads * num_splits * hidden_size, :] _a : List[Any] = (num_heads, num_splits, hidden_size) + input_shape[1:] _a : List[str] = param.view(*UpperCamelCase__ ) _a : int = param.transpose(0 , 1 ).contiguous() _a : Any = param.view(*UpperCamelCase__ ) return param def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' # The converted output model. _a : Any = {} # old versions did not store training args _a : int = input_state_dict.get("""args""" , UpperCamelCase__ ) if ds_args is not None: # do not make the user write a config file when the exact dimensions/sizes are already in the checkpoint # from pprint import pprint # pprint(vars(ds_args)) _a : List[str] = ds_args.padded_vocab_size _a : Tuple = ds_args.max_position_embeddings _a : Union[str, Any] = ds_args.hidden_size _a : List[str] = ds_args.num_layers _a : Any = ds_args.num_attention_heads _a : List[str] = ds_args.ffn_hidden_size # pprint(config) # The number of heads. _a : Dict = config.n_head # The hidden_size per head. _a : Tuple = config.n_embd // config.n_head # Megatron-LM checkpoint version if "checkpoint_version" in input_state_dict.keys(): _a : List[str] = input_state_dict["""checkpoint_version"""] else: _a : str = 0.0 # The model. _a : Tuple = input_state_dict["""model"""] # The language model. _a : Dict = model["""language_model"""] # The embeddings. _a : Tuple = lm["""embedding"""] # The word embeddings. _a : Dict = embeddings["""word_embeddings"""]["""weight"""] # Truncate the embedding table to vocab_size rows. _a : List[Any] = word_embeddings[: config.vocab_size, :] _a : Optional[Any] = word_embeddings # The position embeddings. _a : Any = embeddings["""position_embeddings"""]["""weight"""] # Read the causal mask dimension (seqlen). [max_sequence_length, hidden_size] _a : int = pos_embeddings.size(0 ) if n_positions != config.n_positions: raise ValueError( F"""pos_embeddings.max_sequence_length={n_positions} and config.n_positions={config.n_positions} don't match""" ) # Store the position embeddings. _a : Optional[Any] = pos_embeddings # The transformer. _a : Union[str, Any] = lm["""transformer"""] if """transformer""" in lm.keys() else lm["""encoder"""] # The regex to extract layer names. _a : Optional[Any] = re.compile(R"""layers\.(\d+)\.([a-z0-9_.]+)\.([a-z]+)""" ) # The simple map of names for "automated" rules. _a : List[str] = { """attention.dense""": """.attn.c_proj.""", """self_attention.dense""": """.attn.c_proj.""", """mlp.dense_h_to_4h""": """.mlp.c_fc.""", """mlp.dense_4h_to_h""": """.mlp.c_proj.""", } # Extract the layers. for key, val in transformer.items(): # Match the name. _a : Dict = layer_re.match(UpperCamelCase__ ) # Stop if that's not a layer if m is None: break # The index of the layer. _a : List[str] = int(m.group(1 ) ) # The name of the operation. _a : Optional[int] = m.group(2 ) # Is it a weight or a bias? _a : Optional[Any] = m.group(3 ) # The name of the layer. _a : str = F"""transformer.h.{layer_idx}""" # For layernorm(s), simply store the layer norm. if op_name.endswith("""layernorm""" ): _a : Union[str, Any] = """ln_1""" if op_name.startswith("""input""" ) else """ln_2""" _a : List[Any] = val # Transpose the QKV matrix. elif ( op_name == "attention.query_key_value" or op_name == "self_attention.query_key_value" ) and weight_or_bias == "weight": # Insert a tensor of 1x1xDxD bias. _a : str = torch.tril(torch.ones((n_positions, n_positions) , dtype=torch.floataa ) ).view( 1 , 1 , UpperCamelCase__ , UpperCamelCase__ ) _a : List[str] = causal_mask # Insert a "dummy" tensor for masked_bias. _a : List[str] = torch.tensor(-1e4 , dtype=torch.floataa ) _a : str = masked_bias _a : str = fix_query_key_value_ordering(UpperCamelCase__ , UpperCamelCase__ , 3 , UpperCamelCase__ , UpperCamelCase__ ) # Megatron stores (3*D) x D but transformers-GPT2 expects D x 3*D. _a : Tuple = out_val.transpose(0 , 1 ).contiguous() # Store. _a : Optional[Any] = out_val # Transpose the bias. elif ( op_name == "attention.query_key_value" or op_name == "self_attention.query_key_value" ) and weight_or_bias == "bias": _a : Union[str, Any] = fix_query_key_value_ordering(UpperCamelCase__ , UpperCamelCase__ , 3 , UpperCamelCase__ , UpperCamelCase__ ) # Store. No change of shape. _a : Any = out_val # Transpose the weights. elif weight_or_bias == "weight": _a : int = megatron_to_transformers[op_name] _a : List[str] = val.transpose(0 , 1 ) # Copy the bias. elif weight_or_bias == "bias": _a : int = megatron_to_transformers[op_name] _a : int = val # DEBUG. assert config.n_layer == layer_idx + 1 # The final layernorm. _a : Tuple = transformer["""final_layernorm.weight"""] _a : Dict = transformer["""final_layernorm.bias"""] # For LM head, transformers' wants the matrix to weight embeddings. _a : Union[str, Any] = word_embeddings # It should be done! return output_state_dict def lowerCAmelCase__ ( ): '''simple docstring''' # Create the argument parser. _a : int = argparse.ArgumentParser() parser.add_argument("""--print-checkpoint-structure""" , action="""store_true""" ) parser.add_argument( """path_to_checkpoint""" , type=UpperCamelCase__ , help="""Path to the checkpoint file (.zip archive or direct .pt file)""" , ) parser.add_argument( """--config_file""" , default="""""" , type=UpperCamelCase__ , help="""An optional config json file describing the pre-trained model.""" , ) _a : int = parser.parse_args() # Extract the basename. _a : List[str] = os.path.dirname(args.path_to_checkpoint ) # Load the model. # the .zip is very optional, let's keep it for backward compatibility print(F"""Extracting PyTorch state dictionary from {args.path_to_checkpoint}""" ) if args.path_to_checkpoint.endswith(""".zip""" ): with zipfile.ZipFile(args.path_to_checkpoint , """r""" ) as checkpoint: with checkpoint.open("""release/mp_rank_00/model_optim_rng.pt""" ) as pytorch_dict: _a : Tuple = torch.load(UpperCamelCase__ , map_location="""cpu""" ) else: _a : Dict = torch.load(args.path_to_checkpoint , map_location="""cpu""" ) _a : int = input_state_dict.get("""args""" , UpperCamelCase__ ) # Read the config, or default to the model released by NVIDIA. if args.config_file == "": if ds_args is not None: if ds_args.bias_gelu_fusion: _a : Tuple = """gelu_fast""" elif ds_args.openai_gelu: _a : List[Any] = """gelu_new""" else: _a : Optional[int] = """gelu""" else: # in the very early days this used to be "gelu_new" _a : str = """gelu_new""" # Spell out all parameters in case the defaults change. _a : Any = GPTaConfig( vocab_size=5_0_2_5_7 , n_positions=1_0_2_4 , n_embd=1_0_2_4 , n_layer=2_4 , n_head=1_6 , n_inner=4_0_9_6 , activation_function=UpperCamelCase__ , resid_pdrop=0.1 , embd_pdrop=0.1 , attn_pdrop=0.1 , layer_norm_epsilon=1e-5 , initializer_range=0.02 , summary_type="""cls_index""" , summary_use_proj=UpperCamelCase__ , summary_activation=UpperCamelCase__ , summary_proj_to_labels=UpperCamelCase__ , summary_first_dropout=0.1 , scale_attn_weights=UpperCamelCase__ , use_cache=UpperCamelCase__ , bos_token_id=5_0_2_5_6 , eos_token_id=5_0_2_5_6 , ) else: _a : Optional[Any] = GPTaConfig.from_json_file(args.config_file ) _a : Optional[Any] = ["""GPT2LMHeadModel"""] # Convert. print("""Converting""" ) _a : Dict = convert_megatron_checkpoint(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # Print the structure of converted state dict. if args.print_checkpoint_structure: recursive_print(UpperCamelCase__ , UpperCamelCase__ ) # Add tokenizer class info to config # see https://github.com/huggingface/transformers/issues/13906) if ds_args is not None: _a : Optional[Any] = ds_args.tokenizer_type if tokenizer_type == "GPT2BPETokenizer": _a : Union[str, Any] = """gpt2""" elif tokenizer_type == "PretrainedFromHF": _a : int = ds_args.tokenizer_name_or_path else: raise ValueError(F"""Unrecognized tokenizer_type {tokenizer_type}""" ) else: _a : str = """gpt2""" _a : Tuple = AutoTokenizer.from_pretrained(UpperCamelCase__ ) _a : Any = type(UpperCamelCase__ ).__name__ _a : int = tokenizer_class # Store the config to file. print("""Saving config""" ) config.save_pretrained(UpperCamelCase__ ) # Save tokenizer based on args print(F"""Adding {tokenizer_class} tokenizer files""" ) tokenizer.save_pretrained(UpperCamelCase__ ) # Store the state_dict to file. _a : Optional[Any] = os.path.join(UpperCamelCase__ , """pytorch_model.bin""" ) print(F"""Saving checkpoint to \"{output_checkpoint_file}\"""" ) torch.save(UpperCamelCase__ , UpperCamelCase__ ) #################################################################################################### if __name__ == "__main__": main() ####################################################################################################
294
"""simple docstring""" import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device _snake_case = False class UpperCamelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : Optional[Any] ) -> Optional[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self : Tuple ) -> List[Any]: _a : Tuple = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : List[str] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) _a : Optional[Any] = torch.manual_seed(0 ) _a : Union[str, Any] = pipe.dual_guided( prompt="""first prompt""" , image=UpperCAmelCase__ , text_to_image_strength=0.7_5 , generator=UpperCAmelCase__ , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(UpperCAmelCase__ ) _a : Dict = VersatileDiffusionPipeline.from_pretrained(UpperCAmelCase__ , torch_dtype=torch.floataa ) pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Optional[Any] = generator.manual_seed(0 ) _a : str = pipe.dual_guided( prompt="""first prompt""" , image=UpperCAmelCase__ , text_to_image_strength=0.7_5 , generator=UpperCAmelCase__ , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def _lowercase ( self : Optional[int] ) -> Optional[int]: _a : Optional[int] = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : int = """cyberpunk 2077""" _a : Tuple = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) _a : Tuple = torch.manual_seed(0 ) _a : Any = pipe.dual_guided( prompt=UpperCAmelCase__ , image=UpperCAmelCase__ , text_to_image_strength=0.7_5 , generator=UpperCAmelCase__ , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" , ).images _a : List[str] = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _a : Optional[int] = np.array([0.1_4_4_8, 0.1_6_1_9, 0.1_7_4_1, 0.1_0_8_6, 0.1_1_4_7, 0.1_1_2_8, 0.1_1_9_9, 0.1_1_6_5, 0.1_0_0_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 _a : int = """A painting of a squirrel eating a burger """ _a : Tuple = torch.manual_seed(0 ) _a : Union[str, Any] = pipe.text_to_image( prompt=UpperCAmelCase__ , generator=UpperCAmelCase__ , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" ).images _a : int = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _a : int = np.array([0.3_3_6_7, 0.3_1_6_9, 0.2_6_5_6, 0.3_8_7_0, 0.4_7_9_0, 0.3_7_9_6, 0.4_0_0_9, 0.4_8_7_8, 0.4_7_7_8] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 _a : str = pipe.image_variation(UpperCAmelCase__ , generator=UpperCAmelCase__ , output_type="""numpy""" ).images _a : str = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _a : Optional[Any] = np.array([0.3_0_7_6, 0.3_1_2_3, 0.3_2_8_4, 0.3_7_8_2, 0.3_7_7_0, 0.3_8_9_4, 0.4_2_9_7, 0.4_3_3_1, 0.4_4_5_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1
294
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _snake_case = { 'configuration_vision_encoder_decoder': ['VisionEncoderDecoderConfig', 'VisionEncoderDecoderOnnxConfig'] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ['VisionEncoderDecoderModel'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ['TFVisionEncoderDecoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ['FlaxVisionEncoderDecoderModel'] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys _snake_case = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
294
"""simple docstring""" import json import os from typing import Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = {'vocab_file': 'vocab.json'} _snake_case = { 'vocab_file': { 'mgp-str': 'https://huggingface.co/alibaba-damo/mgp-str-base/blob/main/vocab.json', } } _snake_case = {'mgp-str': 27} class UpperCamelCase ( snake_case_ ): UpperCamelCase : List[str] = VOCAB_FILES_NAMES UpperCamelCase : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : Optional[int] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : List[Any]="[GO]" , UpperCAmelCase__ : Tuple="[GO]" , UpperCAmelCase__ : Optional[int]="[s]" , UpperCAmelCase__ : int="[GO]" , **UpperCAmelCase__ : Dict ) -> int: super().__init__( unk_token=UpperCAmelCase__ , bos_token=UpperCAmelCase__ , eos_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , **UpperCAmelCase__ , ) with open(UpperCAmelCase__ , encoding="""utf-8""" ) as vocab_handle: _a : int = json.load(UpperCAmelCase__ ) _a : Optional[int] = {v: k for k, v in self.vocab.items()} @property def _lowercase ( self : Dict ) -> Union[str, Any]: return len(self.vocab ) def _lowercase ( self : Union[str, Any] ) -> str: return dict(self.vocab , **self.added_tokens_encoder ) def _lowercase ( self : Dict , UpperCAmelCase__ : str ) -> Union[str, Any]: _a : Tuple = [] for s in text: char_tokens.extend(UpperCAmelCase__ ) return char_tokens def _lowercase ( self : List[Any] , UpperCAmelCase__ : str ) -> Dict: return self.vocab.get(UpperCAmelCase__ , self.vocab.get(self.unk_token ) ) def _lowercase ( self : Optional[int] , UpperCAmelCase__ : Tuple ) -> List[Any]: return self.decoder.get(UpperCAmelCase__ ) def _lowercase ( self : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(UpperCAmelCase__ ): logger.error("""Vocabulary path ({}) should be a directory""".format(UpperCAmelCase__ ) ) return _a : Tuple = os.path.join( UpperCAmelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) with open(UpperCAmelCase__ , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.vocab , indent=2 , sort_keys=UpperCAmelCase__ , ensure_ascii=UpperCAmelCase__ ) + """\n""" ) return (vocab_file,)
294
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _snake_case = { 'configuration_distilbert': [ 'DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'DistilBertConfig', 'DistilBertOnnxConfig', ], 'tokenization_distilbert': ['DistilBertTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ['DistilBertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'DistilBertForMaskedLM', 'DistilBertForMultipleChoice', 'DistilBertForQuestionAnswering', 'DistilBertForSequenceClassification', 'DistilBertForTokenClassification', 'DistilBertModel', 'DistilBertPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFDistilBertForMaskedLM', 'TFDistilBertForMultipleChoice', 'TFDistilBertForQuestionAnswering', 'TFDistilBertForSequenceClassification', 'TFDistilBertForTokenClassification', 'TFDistilBertMainLayer', 'TFDistilBertModel', 'TFDistilBertPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'FlaxDistilBertForMaskedLM', 'FlaxDistilBertForMultipleChoice', 'FlaxDistilBertForQuestionAnswering', 'FlaxDistilBertForSequenceClassification', 'FlaxDistilBertForTokenClassification', 'FlaxDistilBertModel', 'FlaxDistilBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_distilbert import ( DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DistilBertConfig, DistilBertOnnxConfig, ) from .tokenization_distilbert import DistilBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_distilbert_fast import DistilBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_distilbert import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, DistilBertPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertMainLayer, TFDistilBertModel, TFDistilBertPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, FlaxDistilBertPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
294
"""simple docstring""" import tempfile import torch from diffusers import IPNDMScheduler from .test_schedulers import SchedulerCommonTest class UpperCamelCase ( snake_case_ ): UpperCamelCase : int = (IPNDMScheduler,) UpperCamelCase : int = (('''num_inference_steps''', 50),) def _lowercase ( self : Union[str, Any] , **UpperCAmelCase__ : Tuple ) -> int: _a : Optional[int] = {"""num_train_timesteps""": 1000} config.update(**UpperCAmelCase__ ) return config def _lowercase ( self : Dict , UpperCAmelCase__ : Any=0 , **UpperCAmelCase__ : Optional[Any] ) -> Union[str, Any]: _a : Optional[int] = dict(self.forward_default_kwargs ) _a : Dict = kwargs.pop("""num_inference_steps""" , UpperCAmelCase__ ) _a : Optional[Any] = self.dummy_sample _a : Union[str, Any] = 0.1 * sample _a : Union[str, Any] = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] for scheduler_class in self.scheduler_classes: _a : Optional[int] = self.get_scheduler_config(**UpperCAmelCase__ ) _a : Union[str, Any] = scheduler_class(**UpperCAmelCase__ ) scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residuals _a : Any = dummy_past_residuals[:] if time_step is None: _a : str = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(UpperCAmelCase__ ) _a : Union[str, Any] = scheduler_class.from_pretrained(UpperCAmelCase__ ) new_scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residuals _a : Optional[Any] = dummy_past_residuals[:] _a : List[Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : str = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" _a : Optional[int] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : Tuple = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def _lowercase ( self : Tuple ) -> List[str]: pass def _lowercase ( self : Optional[int] , UpperCAmelCase__ : List[str]=0 , **UpperCAmelCase__ : Optional[Any] ) -> List[Any]: _a : Optional[Any] = dict(self.forward_default_kwargs ) _a : Optional[Any] = kwargs.pop("""num_inference_steps""" , UpperCAmelCase__ ) _a : Optional[Any] = self.dummy_sample _a : List[Any] = 0.1 * sample _a : Optional[Any] = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] for scheduler_class in self.scheduler_classes: _a : Union[str, Any] = self.get_scheduler_config() _a : Optional[Any] = scheduler_class(**UpperCAmelCase__ ) scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residuals (must be after setting timesteps) _a : Any = dummy_past_residuals[:] if time_step is None: _a : List[Any] = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(UpperCAmelCase__ ) _a : Any = scheduler_class.from_pretrained(UpperCAmelCase__ ) # copy over dummy past residuals new_scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residual (must be after setting timesteps) _a : Optional[Any] = dummy_past_residuals[:] _a : List[str] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : Tuple = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" _a : Union[str, Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : int = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def _lowercase ( self : str , **UpperCAmelCase__ : Any ) -> List[str]: _a : Optional[int] = self.scheduler_classes[0] _a : Optional[Any] = self.get_scheduler_config(**UpperCAmelCase__ ) _a : Union[str, Any] = scheduler_class(**UpperCAmelCase__ ) _a : int = 10 _a : List[Any] = self.dummy_model() _a : str = self.dummy_sample_deter scheduler.set_timesteps(UpperCAmelCase__ ) for i, t in enumerate(scheduler.timesteps ): _a : str = model(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ).prev_sample for i, t in enumerate(scheduler.timesteps ): _a : Union[str, Any] = model(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Any = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ).prev_sample return sample def _lowercase ( self : int ) -> str: _a : Dict = dict(self.forward_default_kwargs ) _a : int = kwargs.pop("""num_inference_steps""" , UpperCAmelCase__ ) for scheduler_class in self.scheduler_classes: _a : Optional[int] = self.get_scheduler_config() _a : Tuple = scheduler_class(**UpperCAmelCase__ ) _a : Tuple = self.dummy_sample _a : Optional[Any] = 0.1 * sample if num_inference_steps is not None and hasattr(UpperCAmelCase__ , """set_timesteps""" ): scheduler.set_timesteps(UpperCAmelCase__ ) elif num_inference_steps is not None and not hasattr(UpperCAmelCase__ , """set_timesteps""" ): _a : List[str] = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) _a : Union[str, Any] = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] _a : Optional[Any] = dummy_past_residuals[:] _a : Optional[Any] = scheduler.timesteps[5] _a : str = scheduler.timesteps[6] _a : Optional[int] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : Union[str, Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) _a : Tuple = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : List[str] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def _lowercase ( self : List[str] ) -> List[str]: for timesteps in [100, 1000]: self.check_over_configs(num_train_timesteps=UpperCAmelCase__ , time_step=UpperCAmelCase__ ) def _lowercase ( self : List[str] ) -> List[str]: for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 100] ): self.check_over_forward(num_inference_steps=UpperCAmelCase__ , time_step=UpperCAmelCase__ ) def _lowercase ( self : int ) -> List[Any]: _a : str = self.full_loop() _a : List[Any] = torch.mean(torch.abs(UpperCAmelCase__ ) ) assert abs(result_mean.item() - 2540529 ) < 10
294
1
"""simple docstring""" from __future__ import annotations import math def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' if num <= 0: _a : Optional[Any] = F"""{num}: Invalid input, please enter a positive integer.""" raise ValueError(UpperCamelCase__ ) _a : Dict = [True] * (num + 1) _a : Optional[Any] = [] _a : Any = 2 _a : Union[str, Any] = int(math.sqrt(UpperCamelCase__ ) ) while start <= end: # If start is a prime if sieve[start] is True: prime.append(UpperCamelCase__ ) # Set multiples of start be False for i in range(start * start , num + 1 , UpperCamelCase__ ): if sieve[i] is True: _a : Optional[int] = False start += 1 for j in range(end + 1 , num + 1 ): if sieve[j] is True: prime.append(UpperCamelCase__ ) return prime if __name__ == "__main__": print(prime_sieve(int(input('Enter a positive integer: ').strip())))
294
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging _snake_case = logging.get_logger(__name__) _snake_case = {'vocab_file': 'spiece.model'} _snake_case = { 'vocab_file': { 'TsinghuaAI/CPM-Generate': 'https://huggingface.co/TsinghuaAI/CPM-Generate/resolve/main/spiece.model', } } class UpperCamelCase ( snake_case_ ): def __init__( self : Tuple , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : int=False , UpperCAmelCase__ : Dict=True , UpperCAmelCase__ : Union[str, Any]=False , UpperCAmelCase__ : Dict="<s>" , UpperCAmelCase__ : Any="</s>" , UpperCAmelCase__ : Any="<unk>" , UpperCAmelCase__ : int="<sep>" , UpperCAmelCase__ : Tuple="<pad>" , UpperCAmelCase__ : Any="<cls>" , UpperCAmelCase__ : Optional[Any]="<mask>" , UpperCAmelCase__ : int=["<eop>", "<eod>"] , UpperCAmelCase__ : Optional[Dict[str, Any]] = None , **UpperCAmelCase__ : List[str] , ) -> None: _a : Optional[int] = AddedToken(UpperCAmelCase__ , lstrip=UpperCAmelCase__ , rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else mask_token _a : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=UpperCAmelCase__ , remove_space=UpperCAmelCase__ , keep_accents=UpperCAmelCase__ , bos_token=UpperCAmelCase__ , eos_token=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , sep_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , cls_token=UpperCAmelCase__ , mask_token=UpperCAmelCase__ , additional_special_tokens=UpperCAmelCase__ , sp_model_kwargs=self.sp_model_kwargs , **UpperCAmelCase__ , ) _a : Optional[Any] = 3 _a : Tuple = do_lower_case _a : Tuple = remove_space _a : Tuple = keep_accents _a : Tuple = vocab_file _a : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCAmelCase__ ) try: import jieba except ModuleNotFoundError as error: raise error.__class__( """You need to install jieba to use CpmTokenizer or CpmTokenizerFast. """ """See https://pypi.org/project/jieba/ for installation.""" ) _a : int = jieba _a : Tuple = str.maketrans(""" \n""" , """\u2582\u2583""" ) @property # Copied from transformers.models.xlnet.tokenization_xlnet.XLNetTokenizer.vocab_size def _lowercase ( self : Optional[Any] ) -> Any: return len(self.sp_model ) def _lowercase ( self : str ) -> Union[str, Any]: _a : int = {self.convert_ids_to_tokens(UpperCAmelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Tuple ) -> List[str]: _a : Tuple = self.__dict__.copy() _a : Tuple = None return state def __setstate__( self : Any , UpperCAmelCase__ : Dict ) -> Dict: _a : Tuple = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): _a : Tuple = {} _a : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _lowercase ( self : List[str] , UpperCAmelCase__ : Union[str, Any] ) -> Dict: if self.remove_space: _a : Optional[int] = """ """.join(inputs.strip().split() ) else: _a : List[Any] = inputs _a : int = outputs.replace("""``""" , """\"""" ).replace("""''""" , """\"""" ) if not self.keep_accents: _a : Optional[Any] = unicodedata.normalize("""NFKD""" , UpperCAmelCase__ ) _a : Dict = """""".join([c for c in outputs if not unicodedata.combining(UpperCAmelCase__ )] ) if self.do_lower_case: _a : Union[str, Any] = outputs.lower() return outputs def _lowercase ( self : List[str] , UpperCAmelCase__ : str ) -> List[str]: _a : str = self.preprocess_text(UpperCAmelCase__ ) _a : Dict = self.sp_model.encode(UpperCAmelCase__ , out_type=UpperCAmelCase__ ) _a : Union[str, Any] = [] for piece in pieces: if len(UpperCAmelCase__ ) > 1 and piece[-1] == str(""",""" ) and piece[-2].isdigit(): _a : Dict = self.sp_model.EncodeAsPieces(piece[:-1].replace(UpperCAmelCase__ , """""" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: _a : Dict = cur_pieces[1:] else: _a : Any = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(UpperCAmelCase__ ) else: new_pieces.append(UpperCAmelCase__ ) return new_pieces def _lowercase ( self : Union[str, Any] , UpperCAmelCase__ : int ) -> int: return self.sp_model.PieceToId(UpperCAmelCase__ ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : Optional[Any] ) -> Any: return self.sp_model.IdToPiece(UpperCAmelCase__ ) def _lowercase ( self : Any , UpperCAmelCase__ : Any ) -> Dict: _a : Dict = """""".join(UpperCAmelCase__ ).replace(UpperCAmelCase__ , """ """ ).strip() return out_string def _lowercase ( self : Any , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ) -> List[int]: _a : Optional[Any] = [self.sep_token_id] _a : Dict = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _lowercase ( self : Tuple , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None , UpperCAmelCase__ : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase__ , token_ids_a=UpperCAmelCase__ , already_has_special_tokens=UpperCAmelCase__ ) if token_ids_a is not None: return ([0] * len(UpperCAmelCase__ )) + [1] + ([0] * len(UpperCAmelCase__ )) + [1, 1] return ([0] * len(UpperCAmelCase__ )) + [1, 1] def _lowercase ( self : str , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ) -> List[int]: _a : Any = [self.sep_token_id] _a : Optional[Any] = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _lowercase ( self : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(UpperCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return _a : Union[str, Any] = os.path.join( UpperCAmelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCAmelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCAmelCase__ , """wb""" ) as fi: _a : Optional[Any] = self.sp_model.serialized_model_proto() fi.write(UpperCAmelCase__ ) return (out_vocab_file,) def _lowercase ( self : Any , *UpperCAmelCase__ : List[str] , **UpperCAmelCase__ : List[str] ) -> List[str]: _a : Tuple = super()._decode(*UpperCAmelCase__ , **UpperCAmelCase__ ) _a : Optional[Any] = text.replace(""" """ , """""" ).replace("""\u2582""" , """ """ ).replace("""\u2583""" , """\n""" ) return text
294
1
"""simple docstring""" def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' # "extended trapezoidal rule" # int(f) = dx/2 * (f1 + 2f2 + ... + fn) _a : str = (boundary[1] - boundary[0]) / steps _a : Union[str, Any] = boundary[0] _a : int = boundary[1] _a : int = make_points(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) _a : str = 0.0 y += (h / 2.0) * f(UpperCamelCase__ ) for i in x_i: # print(i) y += h * f(UpperCamelCase__ ) y += (h / 2.0) * f(UpperCamelCase__ ) return y def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : str = a + h while x < (b - h): yield x _a : List[Any] = x + h def lowerCAmelCase__ ( UpperCamelCase__ ): # enter your function here '''simple docstring''' _a : Dict = (x - 0) * (x - 0) return y def lowerCAmelCase__ ( ): '''simple docstring''' _a : Optional[int] = 0.0 # Lower bound of integration _a : List[str] = 1.0 # Upper bound of integration _a : Optional[int] = 10.0 # define number of steps or resolution _a : Dict = [a, b] # define boundary of integration _a : Union[str, Any] = method_a(UpperCamelCase__ , UpperCamelCase__ ) print(F"""y = {y}""" ) if __name__ == "__main__": main()
294
"""simple docstring""" import inspect import unittest import torch import torch.nn as nn from accelerate.hooks import ( AlignDevicesHook, ModelHook, SequentialHook, add_hook_to_module, attach_align_device_hook, remove_hook_from_module, remove_hook_from_submodules, ) from accelerate.test_utils import require_multi_gpu class UpperCamelCase ( nn.Module ): def __init__( self : Union[str, Any] ) -> int: super().__init__() _a : Optional[Any] = nn.Linear(3 , 4 ) _a : Tuple = nn.BatchNormad(4 ) _a : Dict = nn.Linear(4 , 5 ) def _lowercase ( self : Optional[int] , UpperCAmelCase__ : List[str] ) -> int: return self.lineara(self.batchnorm(self.lineara(UpperCAmelCase__ ) ) ) class UpperCamelCase ( snake_case_ ): def _lowercase ( self : Any , UpperCAmelCase__ : Any , *UpperCAmelCase__ : List[str] , **UpperCAmelCase__ : Optional[int] ) -> Optional[int]: return (args[0] + 1,) + args[1:], kwargs class UpperCamelCase ( snake_case_ ): def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Optional[Any] ) -> List[str]: return output + 1 class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : Dict ) -> str: _a : List[Any] = ModelForTest() _a : str = ModelHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) self.assertEqual(test_model._hf_hook , UpperCAmelCase__ ) self.assertTrue(hasattr(UpperCAmelCase__ , """_old_forward""" ) ) # Check adding the hook did not change the name or the signature self.assertEqual(test_model.forward.__name__ , """forward""" ) self.assertListEqual(list(inspect.signature(test_model.forward ).parameters ) , ["""x"""] ) remove_hook_from_module(UpperCAmelCase__ ) self.assertFalse(hasattr(UpperCAmelCase__ , """_hf_hook""" ) ) self.assertFalse(hasattr(UpperCAmelCase__ , """_old_forward""" ) ) def _lowercase ( self : Optional[int] ) -> Optional[int]: _a : Dict = ModelForTest() _a : Dict = ModelHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ , append=UpperCAmelCase__ ) self.assertEqual(isinstance(test_model._hf_hook , UpperCAmelCase__ ) , UpperCAmelCase__ ) self.assertEqual(len(test_model._hf_hook.hooks ) , 2 ) self.assertTrue(hasattr(UpperCAmelCase__ , """_old_forward""" ) ) # Check adding the hook did not change the name or the signature self.assertEqual(test_model.forward.__name__ , """forward""" ) self.assertListEqual(list(inspect.signature(test_model.forward ).parameters ) , ["""x"""] ) remove_hook_from_module(UpperCAmelCase__ ) self.assertFalse(hasattr(UpperCAmelCase__ , """_hf_hook""" ) ) self.assertFalse(hasattr(UpperCAmelCase__ , """_old_forward""" ) ) def _lowercase ( self : Dict ) -> int: _a : str = ModelForTest() _a : List[Any] = torch.randn(2 , 3 ) _a : Optional[Any] = test_model(x + 1 ) _a : str = test_model(x + 2 ) _a : Union[str, Any] = PreForwardHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Tuple = test_model(UpperCAmelCase__ ) self.assertTrue(torch.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1E-5 ) ) # Attaching a hook to a model when it already has one replaces, does not chain _a : int = PreForwardHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : str = test_model(UpperCAmelCase__ ) self.assertTrue(torch.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1E-5 ) ) # You need to use the sequential hook to chain two or more hooks _a : int = SequentialHook(PreForwardHook() , PreForwardHook() ) add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Tuple = test_model(UpperCAmelCase__ ) assert torch.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1E-5 ) def _lowercase ( self : Tuple ) -> int: _a : Tuple = ModelForTest() _a : Union[str, Any] = torch.randn(2 , 3 ) _a : Optional[int] = test_model(UpperCAmelCase__ ) _a : int = PostForwardHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[str] = test_model(UpperCAmelCase__ ) self.assertTrue(torch.allclose(UpperCAmelCase__ , output + 1 , atol=1E-5 ) ) # Attaching a hook to a model when it already has one replaces, does not chain _a : List[Any] = PostForwardHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Dict = test_model(UpperCAmelCase__ ) self.assertTrue(torch.allclose(UpperCAmelCase__ , output + 1 , atol=1E-5 ) ) # You need to use the sequential hook to chain two or more hooks _a : Any = SequentialHook(PostForwardHook() , PostForwardHook() ) add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Optional[int] = test_model(UpperCAmelCase__ ) assert torch.allclose(UpperCAmelCase__ , output + 2 , atol=1E-5 ) def _lowercase ( self : Dict ) -> Optional[Any]: _a : Any = ModelForTest() _a : List[Any] = torch.randn(2 , 3 ) _a : Dict = test_model(UpperCAmelCase__ ) _a : Any = PostForwardHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[str] = test_model(UpperCAmelCase__ ) self.assertTrue(torch.allclose(UpperCAmelCase__ , output + 1 ) ) self.assertTrue(outputa.requires_grad ) _a : Any = True _a : Union[str, Any] = test_model(UpperCAmelCase__ ) self.assertFalse(outputa.requires_grad ) @require_multi_gpu def _lowercase ( self : Optional[Any] ) -> str: _a : List[Any] = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices add_hook_to_module(model.lineara , AlignDevicesHook(execution_device=0 ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(execution_device=0 ) ) add_hook_to_module(model.lineara , AlignDevicesHook(execution_device=1 ) ) self.assertEqual(model.lineara.weight.device , torch.device(0 ) ) self.assertEqual(model.batchnorm.weight.device , torch.device(0 ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device(0 ) ) self.assertEqual(model.lineara.weight.device , torch.device(1 ) ) # We can still make a forward pass. The input does not need to be on any particular device _a : Optional[int] = torch.randn(2 , 3 ) _a : Any = model(UpperCAmelCase__ ) self.assertEqual(output.device , torch.device(1 ) ) # We can add a general hook to put back output on same device as input. add_hook_to_module(UpperCAmelCase__ , AlignDevicesHook(io_same_device=UpperCAmelCase__ ) ) _a : str = torch.randn(2 , 3 ).to(0 ) _a : Union[str, Any] = model(UpperCAmelCase__ ) self.assertEqual(output.device , torch.device(0 ) ) def _lowercase ( self : str ) -> Union[str, Any]: _a : int = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices _a : List[Any] = {"""execution_device""": 0 if torch.cuda.is_available() else """cpu""", """offload""": True} add_hook_to_module(model.lineara , AlignDevicesHook(**UpperCAmelCase__ ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(**UpperCAmelCase__ ) ) add_hook_to_module(model.lineara , AlignDevicesHook(**UpperCAmelCase__ ) ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) # Buffers are not included in the offload by default, so are on the execution device _a : Dict = torch.device(hook_kwargs["""execution_device"""] ) self.assertEqual(model.batchnorm.running_mean.device , UpperCAmelCase__ ) _a : int = torch.randn(2 , 3 ) _a : str = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_module(model.lineara ) remove_hook_from_module(model.batchnorm ) remove_hook_from_module(model.lineara ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # Now test with buffers included in the offload _a : List[str] = { """execution_device""": 0 if torch.cuda.is_available() else """cpu""", """offload""": True, """offload_buffers""": True, } add_hook_to_module(model.lineara , AlignDevicesHook(**UpperCAmelCase__ ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(**UpperCAmelCase__ ) ) add_hook_to_module(model.lineara , AlignDevicesHook(**UpperCAmelCase__ ) ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device("""meta""" ) ) _a : Tuple = torch.randn(2 , 3 ) _a : Union[str, Any] = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_module(model.lineara ) remove_hook_from_module(model.batchnorm ) remove_hook_from_module(model.lineara ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) def _lowercase ( self : Tuple ) -> List[str]: _a : str = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices _a : Union[str, Any] = 0 if torch.cuda.is_available() else """cpu""" attach_align_device_hook(UpperCAmelCase__ , execution_device=UpperCAmelCase__ , offload=UpperCAmelCase__ ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) # Buffers are not included in the offload by default, so are on the execution device _a : Dict = torch.device(UpperCAmelCase__ ) self.assertEqual(model.batchnorm.running_mean.device , UpperCAmelCase__ ) _a : Union[str, Any] = torch.randn(2 , 3 ) _a : List[Any] = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(UpperCAmelCase__ ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # Now test with buffers included in the offload attach_align_device_hook(UpperCAmelCase__ , execution_device=UpperCAmelCase__ , offload=UpperCAmelCase__ , offload_buffers=UpperCAmelCase__ ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device("""meta""" ) ) _a : List[str] = torch.randn(2 , 3 ) _a : Union[str, Any] = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(UpperCAmelCase__ ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) def _lowercase ( self : Dict ) -> str: _a : Optional[Any] = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices _a : str = 0 if torch.cuda.is_available() else """cpu""" attach_align_device_hook( UpperCAmelCase__ , execution_device=UpperCAmelCase__ , offload=UpperCAmelCase__ , weights_map=model.state_dict() ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) # Buffers are not included in the offload by default, so are on the execution device _a : Union[str, Any] = torch.device(UpperCAmelCase__ ) self.assertEqual(model.batchnorm.running_mean.device , UpperCAmelCase__ ) _a : Union[str, Any] = torch.randn(2 , 3 ) _a : int = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(UpperCAmelCase__ ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # Now test with buffers included in the offload attach_align_device_hook( UpperCAmelCase__ , execution_device=UpperCAmelCase__ , offload=UpperCAmelCase__ , weights_map=model.state_dict() , offload_buffers=UpperCAmelCase__ , ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device("""meta""" ) ) _a : Any = torch.randn(2 , 3 ) _a : int = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(UpperCAmelCase__ ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) )
294
1
"""simple docstring""" import shutil import tempfile import unittest from transformers import ClapFeatureExtractor, ClapProcessor, RobertaTokenizer, RobertaTokenizerFast from transformers.testing_utils import require_sentencepiece, require_torchaudio from .test_feature_extraction_clap import floats_list @require_torchaudio @require_sentencepiece class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : int ) -> List[str]: _a : Any = """laion/clap-htsat-unfused""" _a : Union[str, Any] = tempfile.mkdtemp() def _lowercase ( self : List[Any] , **UpperCAmelCase__ : Any ) -> Dict: return RobertaTokenizer.from_pretrained(self.checkpoint , **UpperCAmelCase__ ) def _lowercase ( self : List[Any] , **UpperCAmelCase__ : List[str] ) -> int: return ClapFeatureExtractor.from_pretrained(self.checkpoint , **UpperCAmelCase__ ) def _lowercase ( self : List[Any] ) -> Tuple: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : List[str] ) -> Optional[int]: _a : List[str] = self.get_tokenizer() _a : Any = self.get_feature_extractor() _a : Optional[Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) _a : List[str] = ClapProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCAmelCase__ ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , UpperCAmelCase__ ) def _lowercase ( self : Dict ) -> Optional[int]: _a : Tuple = ClapProcessor(tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() ) processor.save_pretrained(self.tmpdirname ) _a : Dict = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) _a : Union[str, Any] = self.get_feature_extractor(do_normalize=UpperCAmelCase__ , padding_value=1.0 ) _a : Union[str, Any] = ClapProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=UpperCAmelCase__ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCAmelCase__ ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.feature_extractor , UpperCAmelCase__ ) def _lowercase ( self : List[str] ) -> Optional[Any]: _a : Optional[int] = self.get_feature_extractor() _a : Tuple = self.get_tokenizer() _a : List[Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) _a : Any = floats_list((3, 1000) ) _a : List[Any] = feature_extractor(UpperCAmelCase__ , return_tensors="""np""" ) _a : List[str] = processor(audios=UpperCAmelCase__ , return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def _lowercase ( self : Tuple ) -> Optional[int]: _a : List[str] = self.get_feature_extractor() _a : Any = self.get_tokenizer() _a : Any = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) _a : Optional[int] = """This is a test string""" _a : Tuple = processor(text=UpperCAmelCase__ ) _a : int = tokenizer(UpperCAmelCase__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _lowercase ( self : List[Any] ) -> Any: _a : str = self.get_feature_extractor() _a : List[str] = self.get_tokenizer() _a : List[Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) _a : Any = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _a : Dict = processor.batch_decode(UpperCAmelCase__ ) _a : Any = tokenizer.batch_decode(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def _lowercase ( self : Dict ) -> List[str]: _a : str = self.get_feature_extractor() _a : Optional[Any] = self.get_tokenizer() _a : Union[str, Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) self.assertListEqual( processor.model_input_names[2:] , feature_extractor.model_input_names , msg="""`processor` and `feature_extractor` model input names do not match""" , )
294
"""simple docstring""" from __future__ import annotations def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' print(F"""Vertex\tShortest Distance from vertex {src}""" ) for i, d in enumerate(UpperCamelCase__ ): print(F"""{i}\t\t{d}""" ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' for j in range(UpperCamelCase__ ): _a , _a , _a : List[str] = (graph[j][k] for k in ["""src""", """dst""", """weight"""]) if distance[u] != float("""inf""" ) and distance[u] + w < distance[v]: return True return False def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Dict = [float("""inf""" )] * vertex_count _a : Any = 0.0 for _ in range(vertex_count - 1 ): for j in range(UpperCamelCase__ ): _a , _a , _a : List[Any] = (graph[j][k] for k in ["""src""", """dst""", """weight"""]) if distance[u] != float("""inf""" ) and distance[u] + w < distance[v]: _a : Any = distance[u] + w _a : Union[str, Any] = check_negative_cycle(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) if negative_cycle_exists: raise Exception("""Negative cycle found""" ) return distance if __name__ == "__main__": import doctest doctest.testmod() _snake_case = int(input('Enter number of vertices: ').strip()) _snake_case = int(input('Enter number of edges: ').strip()) _snake_case = [{} for _ in range(E)] for i in range(E): print('Edge ', i + 1) _snake_case , _snake_case , _snake_case = ( int(x) for x in input('Enter source, destination, weight: ').strip().split(' ') ) _snake_case = {'src': src, 'dst': dest, 'weight': weight} _snake_case = int(input('\nEnter shortest path source:').strip()) _snake_case = bellman_ford(graph, V, E, source) print_distance(shortest_distance, 0)
294
1
"""simple docstring""" import logging import os from typing import List, TextIO, Union from conllu import parse_incr from utils_ner import InputExample, Split, TokenClassificationTask _snake_case = logging.getLogger(__name__) class UpperCamelCase ( snake_case_ ): def __init__( self : Optional[Any] , UpperCAmelCase__ : Optional[int]=-1 ) -> Tuple: # in NER datasets, the last column is usually reserved for NER label _a : Optional[int] = label_idx def _lowercase ( self : Any , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Union[Split, str] ) -> List[InputExample]: if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): _a : Any = mode.value _a : Optional[int] = os.path.join(UpperCAmelCase__ , f"""{mode}.txt""" ) _a : int = 1 _a : int = [] with open(UpperCAmelCase__ , encoding="""utf-8""" ) as f: _a : str = [] _a : str = [] for line in f: if line.startswith("""-DOCSTART-""" ) or line == "" or line == "\n": if words: examples.append(InputExample(guid=f"""{mode}-{guid_index}""" , words=UpperCAmelCase__ , labels=UpperCAmelCase__ ) ) guid_index += 1 _a : List[str] = [] _a : str = [] else: _a : List[Any] = line.split(""" """ ) words.append(splits[0] ) if len(UpperCAmelCase__ ) > 1: labels.append(splits[self.label_idx].replace("""\n""" , """""" ) ) else: # Examples could have no label for mode = "test" labels.append("""O""" ) if words: examples.append(InputExample(guid=f"""{mode}-{guid_index}""" , words=UpperCAmelCase__ , labels=UpperCAmelCase__ ) ) return examples def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : List ) -> Union[str, Any]: _a : List[str] = 0 for line in test_input_reader: if line.startswith("""-DOCSTART-""" ) or line == "" or line == "\n": writer.write(UpperCAmelCase__ ) if not preds_list[example_id]: example_id += 1 elif preds_list[example_id]: _a : int = line.split()[0] + """ """ + preds_list[example_id].pop(0 ) + """\n""" writer.write(UpperCAmelCase__ ) else: logger.warning("""Maximum sequence length exceeded: No prediction for '%s'.""" , line.split()[0] ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : str ) -> List[str]: if path: with open(UpperCAmelCase__ , """r""" ) as f: _a : List[Any] = f.read().splitlines() if "O" not in labels: _a : Union[str, Any] = ["""O"""] + labels return labels else: return ["O", "B-MISC", "I-MISC", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC"] class UpperCamelCase ( snake_case_ ): def __init__( self : Union[str, Any] ) -> List[str]: # in CONLL2003 dataset chunk column is second-to-last super().__init__(label_idx=-2 ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : str ) -> List[str]: if path: with open(UpperCAmelCase__ , """r""" ) as f: _a : Optional[int] = f.read().splitlines() if "O" not in labels: _a : Optional[Any] = ["""O"""] + labels return labels else: return [ "O", "B-ADVP", "B-INTJ", "B-LST", "B-PRT", "B-NP", "B-SBAR", "B-VP", "B-ADJP", "B-CONJP", "B-PP", "I-ADVP", "I-INTJ", "I-LST", "I-PRT", "I-NP", "I-SBAR", "I-VP", "I-ADJP", "I-CONJP", "I-PP", ] class UpperCamelCase ( snake_case_ ): def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Union[Split, str] ) -> List[InputExample]: if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): _a : List[Any] = mode.value _a : Union[str, Any] = os.path.join(UpperCAmelCase__ , f"""{mode}.txt""" ) _a : List[str] = 1 _a : Optional[Any] = [] with open(UpperCAmelCase__ , encoding="""utf-8""" ) as f: for sentence in parse_incr(UpperCAmelCase__ ): _a : List[Any] = [] _a : Any = [] for token in sentence: words.append(token["""form"""] ) labels.append(token["""upos"""] ) assert len(UpperCAmelCase__ ) == len(UpperCAmelCase__ ) if words: examples.append(InputExample(guid=f"""{mode}-{guid_index}""" , words=UpperCAmelCase__ , labels=UpperCAmelCase__ ) ) guid_index += 1 return examples def _lowercase ( self : Tuple , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : List ) -> Dict: _a : Optional[Any] = 0 for sentence in parse_incr(UpperCAmelCase__ ): _a : List[str] = preds_list[example_id] _a : str = """""" for token in sentence: out += f"""{token['form']} ({token['upos']}|{s_p.pop(0 )}) """ out += "\n" writer.write(UpperCAmelCase__ ) example_id += 1 def _lowercase ( self : List[str] , UpperCAmelCase__ : str ) -> List[str]: if path: with open(UpperCAmelCase__ , """r""" ) as f: return f.read().splitlines() else: return [ "ADJ", "ADP", "ADV", "AUX", "CCONJ", "DET", "INTJ", "NOUN", "NUM", "PART", "PRON", "PROPN", "PUNCT", "SCONJ", "SYM", "VERB", "X", ]
294
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _snake_case = { 'configuration_transfo_xl': ['TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TransfoXLConfig'], 'tokenization_transfo_xl': ['TransfoXLCorpus', 'TransfoXLTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST', 'AdaptiveEmbedding', 'TransfoXLForSequenceClassification', 'TransfoXLLMHeadModel', 'TransfoXLModel', 'TransfoXLPreTrainedModel', 'load_tf_weights_in_transfo_xl', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFAdaptiveEmbedding', 'TFTransfoXLForSequenceClassification', 'TFTransfoXLLMHeadModel', 'TFTransfoXLMainLayer', 'TFTransfoXLModel', 'TFTransfoXLPreTrainedModel', ] if TYPE_CHECKING: from .configuration_transfo_xl import TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, TransfoXLConfig from .tokenization_transfo_xl import TransfoXLCorpus, TransfoXLTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_transfo_xl import ( TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, AdaptiveEmbedding, TransfoXLForSequenceClassification, TransfoXLLMHeadModel, TransfoXLModel, TransfoXLPreTrainedModel, load_tf_weights_in_transfo_xl, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_transfo_xl import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFAdaptiveEmbedding, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLMainLayer, TFTransfoXLModel, TFTransfoXLPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
294
1
"""simple docstring""" import argparse import torch from transformers import LxmertConfig, LxmertForPreTraining, load_tf_weights_in_lxmert from transformers.utils import logging logging.set_verbosity_info() def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' # Initialise PyTorch model _a : Union[str, Any] = LxmertConfig.from_json_file(UpperCamelCase__ ) print(F"""Building PyTorch model from configuration: {config}""" ) _a : Union[str, Any] = LxmertForPreTraining(UpperCamelCase__ ) # Load weights from tf checkpoint load_tf_weights_in_lxmert(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # Save pytorch-model print(F"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , UpperCamelCase__ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help='The config json file corresponding to the pre-trained model. \nThis specifies the model architecture.', ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) _snake_case = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
294
"""simple docstring""" def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , ): '''simple docstring''' _a : Optional[Any] = [redshift, radiation_density, matter_density, dark_energy] if any(p < 0 for p in parameters ): raise ValueError("""All input parameters must be positive""" ) if any(p > 1 for p in parameters[1:4] ): raise ValueError("""Relative densities cannot be greater than one""" ) else: _a : Tuple = 1 - (matter_density + radiation_density + dark_energy) _a : int = ( radiation_density * (redshift + 1) ** 4 + matter_density * (redshift + 1) ** 3 + curvature * (redshift + 1) ** 2 + dark_energy ) _a : List[str] = hubble_constant * e_a ** (1 / 2) return hubble if __name__ == "__main__": import doctest # run doctest doctest.testmod() # demo LCDM approximation _snake_case = 0.3 print( hubble_parameter( hubble_constant=68.3, radiation_density=1e-4, matter_density=matter_density, dark_energy=1 - matter_density, redshift=0, ) )
294
1
"""simple docstring""" from __future__ import annotations def lowerCAmelCase__ ( UpperCamelCase__ = 4 ): '''simple docstring''' _a : Tuple = abs(UpperCamelCase__ ) or 4 return [[1 + x + y * row_size for x in range(UpperCamelCase__ )] for y in range(UpperCamelCase__ )] def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' return reverse_row(transpose(UpperCamelCase__ ) ) # OR.. transpose(reverse_column(matrix)) def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' return reverse_row(reverse_column(UpperCamelCase__ ) ) # OR.. reverse_column(reverse_row(matrix)) def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' return reverse_column(transpose(UpperCamelCase__ ) ) # OR.. transpose(reverse_row(matrix)) def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : Optional[Any] = [list(UpperCamelCase__ ) for x in zip(*UpperCamelCase__ )] return matrix def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : List[str] = matrix[::-1] return matrix def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : Any = [x[::-1] for x in matrix] return matrix def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' for i in matrix: print(*UpperCamelCase__ ) if __name__ == "__main__": _snake_case = make_matrix() print('\norigin:\n') print_matrix(matrix) print('\nrotate 90 counterclockwise:\n') print_matrix(rotate_aa(matrix)) _snake_case = make_matrix() print('\norigin:\n') print_matrix(matrix) print('\nrotate 180:\n') print_matrix(rotate_aaa(matrix)) _snake_case = make_matrix() print('\norigin:\n') print_matrix(matrix) print('\nrotate 270 counterclockwise:\n') print_matrix(rotate_aaa(matrix))
294
"""simple docstring""" import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : List[Any] ) -> Dict: _a : Optional[int] = tempfile.mkdtemp() _a : Optional[Any] = SamImageProcessor() _a : int = SamProcessor(UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) def _lowercase ( self : Tuple , **UpperCAmelCase__ : Any ) -> Any: return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase__ ).image_processor def _lowercase ( self : str ) -> int: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Tuple ) -> Dict: _a : List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _a : Optional[int] = [Image.fromarray(np.moveaxis(UpperCAmelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def _lowercase ( self : Dict ) -> Dict: _a : List[Any] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _a : Tuple = self.get_image_processor(do_normalize=UpperCAmelCase__ , padding_value=1.0 ) _a : Tuple = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=UpperCAmelCase__ , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCAmelCase__ ) def _lowercase ( self : Union[str, Any] ) -> Tuple: _a : Optional[Any] = self.get_image_processor() _a : int = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Union[str, Any] = self.prepare_image_inputs() _a : List[str] = image_processor(UpperCAmelCase__ , return_tensors="""np""" ) _a : List[str] = processor(images=UpperCAmelCase__ , return_tensors="""np""" ) input_feat_extract.pop("""original_sizes""" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("""reshaped_input_sizes""" ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_torch def _lowercase ( self : Optional[Any] ) -> Optional[Any]: _a : Optional[Any] = self.get_image_processor() _a : Dict = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Tuple = [torch.ones((1, 3, 5, 5) )] _a : Tuple = [[1764, 2646]] _a : Optional[int] = [[683, 1024]] _a : List[Any] = processor.post_process_masks(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) _a : int = processor.post_process_masks( UpperCAmelCase__ , torch.tensor(UpperCAmelCase__ ) , torch.tensor(UpperCAmelCase__ ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np _a : Optional[Any] = [np.ones((1, 3, 5, 5) )] _a : Tuple = processor.post_process_masks(UpperCAmelCase__ , np.array(UpperCAmelCase__ ) , np.array(UpperCAmelCase__ ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) _a : List[str] = [[1, 0], [0, 1]] with self.assertRaises(UpperCAmelCase__ ): _a : str = processor.post_process_masks(UpperCAmelCase__ , np.array(UpperCAmelCase__ ) , np.array(UpperCAmelCase__ ) ) @require_vision @require_tf class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : Any ) -> List[str]: _a : List[str] = tempfile.mkdtemp() _a : Any = SamImageProcessor() _a : Union[str, Any] = SamProcessor(UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) def _lowercase ( self : List[str] , **UpperCAmelCase__ : Any ) -> List[str]: return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase__ ).image_processor def _lowercase ( self : Optional[Any] ) -> Union[str, Any]: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Dict ) -> List[str]: _a : List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _a : List[Any] = [Image.fromarray(np.moveaxis(UpperCAmelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def _lowercase ( self : Union[str, Any] ) -> Union[str, Any]: _a : Optional[int] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _a : str = self.get_image_processor(do_normalize=UpperCAmelCase__ , padding_value=1.0 ) _a : Union[str, Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=UpperCAmelCase__ , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCAmelCase__ ) def _lowercase ( self : List[Any] ) -> str: _a : Union[str, Any] = self.get_image_processor() _a : Dict = SamProcessor(image_processor=UpperCAmelCase__ ) _a : int = self.prepare_image_inputs() _a : List[str] = image_processor(UpperCAmelCase__ , return_tensors="""np""" ) _a : List[str] = processor(images=UpperCAmelCase__ , return_tensors="""np""" ) input_feat_extract.pop("""original_sizes""" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("""reshaped_input_sizes""" ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_tf def _lowercase ( self : Optional[Any] ) -> int: _a : Optional[Any] = self.get_image_processor() _a : Dict = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Any = [tf.ones((1, 3, 5, 5) )] _a : Tuple = [[1764, 2646]] _a : str = [[683, 1024]] _a : Union[str, Any] = processor.post_process_masks(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , return_tensors="""tf""" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) _a : Union[str, Any] = processor.post_process_masks( UpperCAmelCase__ , tf.convert_to_tensor(UpperCAmelCase__ ) , tf.convert_to_tensor(UpperCAmelCase__ ) , return_tensors="""tf""" , ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np _a : List[Any] = [np.ones((1, 3, 5, 5) )] _a : Optional[int] = processor.post_process_masks( UpperCAmelCase__ , np.array(UpperCAmelCase__ ) , np.array(UpperCAmelCase__ ) , return_tensors="""tf""" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) _a : Dict = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): _a : List[Any] = processor.post_process_masks( UpperCAmelCase__ , np.array(UpperCAmelCase__ ) , np.array(UpperCAmelCase__ ) , return_tensors="""tf""" ) @require_vision @require_torchvision class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : str ) -> Optional[Any]: _a : Optional[Any] = tempfile.mkdtemp() _a : Dict = SamImageProcessor() _a : List[str] = SamProcessor(UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) def _lowercase ( self : Any , **UpperCAmelCase__ : Dict ) -> int: return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase__ ).image_processor def _lowercase ( self : Tuple ) -> List[Any]: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : str ) -> int: _a : str = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _a : int = [Image.fromarray(np.moveaxis(UpperCAmelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def _lowercase ( self : int ) -> List[Any]: _a : Optional[Any] = self.get_image_processor() _a : Optional[Any] = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Tuple = np.random.randint(0 , 2 , size=(1, 3, 5, 5) ).astype(np.floataa ) _a : str = [tf.convert_to_tensor(UpperCAmelCase__ )] _a : Optional[int] = [torch.tensor(UpperCAmelCase__ )] _a : Union[str, Any] = [[1764, 2646]] _a : List[str] = [[683, 1024]] _a : Optional[int] = processor.post_process_masks( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , return_tensors="""tf""" ) _a : List[str] = processor.post_process_masks( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , return_tensors="""pt""" ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def _lowercase ( self : str ) -> Optional[Any]: _a : List[Any] = self.get_image_processor() _a : Any = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Dict = self.prepare_image_inputs() _a : List[str] = image_processor(UpperCAmelCase__ , return_tensors="""pt""" )["""pixel_values"""].numpy() _a : str = processor(images=UpperCAmelCase__ , return_tensors="""pt""" )["""pixel_values"""].numpy() _a : Optional[Any] = image_processor(UpperCAmelCase__ , return_tensors="""tf""" )["""pixel_values"""].numpy() _a : Optional[int] = processor(images=UpperCAmelCase__ , return_tensors="""tf""" )["""pixel_values"""].numpy() self.assertTrue(np.allclose(UpperCAmelCase__ , UpperCAmelCase__ ) ) self.assertTrue(np.allclose(UpperCAmelCase__ , UpperCAmelCase__ ) ) self.assertTrue(np.allclose(UpperCAmelCase__ , UpperCAmelCase__ ) )
294
1
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require_tokenizers class UpperCamelCase ( unittest.TestCase ): @slow def _lowercase ( self : Any ) -> int: _a : List[Any] = AutoModelForSeqaSeqLM.from_pretrained("""google/mt5-small""" , return_dict=UpperCAmelCase__ ).to(UpperCAmelCase__ ) _a : int = AutoTokenizer.from_pretrained("""google/mt5-small""" ) _a : Optional[int] = tokenizer("""Hello there""" , return_tensors="""pt""" ).input_ids _a : Union[str, Any] = tokenizer("""Hi I am""" , return_tensors="""pt""" ).input_ids _a : str = model(input_ids.to(UpperCAmelCase__ ) , labels=labels.to(UpperCAmelCase__ ) ).loss _a : Union[str, Any] = -(labels.shape[-1] * loss.item()) _a : Union[str, Any] = -8_4.9_1_2_7 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
294
"""simple docstring""" import argparse import gc import json import os import re import torch from huggingface_hub import hf_hub_download from transformers import AutoModelForCausalLM, AutoTokenizer, PreTrainedTokenizerFast, RwkvConfig from transformers.modeling_utils import WEIGHTS_INDEX_NAME, shard_checkpoint _snake_case = { '169M': 12, '430M': 24, '1B5': 24, '3B': 32, '7B': 32, '14B': 40, } _snake_case = { '169M': 768, '430M': 1024, '1B5': 2048, '3B': 2560, '7B': 4096, '14B': 5120, } def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : int = list(state_dict.keys() ) for name in state_dict_keys: _a : str = state_dict.pop(UpperCamelCase__ ) # emb -> embedding if name.startswith("""emb.""" ): _a : Dict = name.replace("""emb.""" , """embeddings.""" ) # ln_0 -> pre_ln (only present at block 0) if name.startswith("""blocks.0.ln0""" ): _a : Dict = name.replace("""blocks.0.ln0""" , """blocks.0.pre_ln""" ) # att -> attention _a : Any = re.sub(R"""blocks\.(\d+)\.att""" , R"""blocks.\1.attention""" , UpperCamelCase__ ) # ffn -> feed_forward _a : int = re.sub(R"""blocks\.(\d+)\.ffn""" , R"""blocks.\1.feed_forward""" , UpperCamelCase__ ) # time_mix_k -> time_mix_key and reshape if name.endswith(""".time_mix_k""" ): _a : List[str] = name.replace(""".time_mix_k""" , """.time_mix_key""" ) # time_mix_v -> time_mix_value and reshape if name.endswith(""".time_mix_v""" ): _a : Tuple = name.replace(""".time_mix_v""" , """.time_mix_value""" ) # time_mix_r -> time_mix_key and reshape if name.endswith(""".time_mix_r""" ): _a : Dict = name.replace(""".time_mix_r""" , """.time_mix_receptance""" ) if name != "head.weight": _a : Optional[int] = """rwkv.""" + name _a : Any = weight return state_dict def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=False , UpperCamelCase__=None ): '''simple docstring''' # 1. If possible, build the tokenizer. if tokenizer_file is None: print("""No `--tokenizer_file` provided, we will use the default tokenizer.""" ) _a : Tuple = 5_0_2_7_7 _a : str = AutoTokenizer.from_pretrained("""EleutherAI/gpt-neox-20b""" ) else: _a : int = PreTrainedTokenizerFast(tokenizer_file=UpperCamelCase__ ) _a : int = len(UpperCamelCase__ ) tokenizer.save_pretrained(UpperCamelCase__ ) # 2. Build the config _a : Optional[Any] = list(NUM_HIDDEN_LAYERS_MAPPING.keys() ) if size is None: # Try to infer size from the checkpoint name for candidate in possible_sizes: if candidate in checkpoint_file: _a : Tuple = candidate break if size is None: raise ValueError("""Could not infer the size, please provide it with the `--size` argument.""" ) if size not in possible_sizes: raise ValueError(F"""`size` should be one of {possible_sizes}, got {size}.""" ) _a : List[Any] = RwkvConfig( vocab_size=UpperCamelCase__ , num_hidden_layers=NUM_HIDDEN_LAYERS_MAPPING[size] , hidden_size=HIDEN_SIZE_MAPPING[size] , ) config.save_pretrained(UpperCamelCase__ ) # 3. Download model file then convert state_dict _a : str = hf_hub_download(UpperCamelCase__ , UpperCamelCase__ ) _a : int = torch.load(UpperCamelCase__ , map_location="""cpu""" ) _a : List[str] = convert_state_dict(UpperCamelCase__ ) # 4. Split in shards and save _a , _a : List[str] = shard_checkpoint(UpperCamelCase__ ) for shard_file, shard in shards.items(): torch.save(UpperCamelCase__ , os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) if index is not None: _a : Optional[Any] = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) # Save the index as well with open(UpperCamelCase__ , """w""" , encoding="""utf-8""" ) as f: _a : Dict = json.dumps(UpperCamelCase__ , indent=2 , sort_keys=UpperCamelCase__ ) + """\n""" f.write(UpperCamelCase__ ) # 5. Clean up shards (for some reason the file PyTorch saves take the same space as the whole state_dict print( """Cleaning up shards. This may error with an OOM error, it this is the case don't worry you still have converted the model.""" ) _a : List[Any] = list(shards.keys() ) del state_dict del shards gc.collect() for shard_file in shard_files: _a : Any = torch.load(os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) torch.save({k: v.cpu().clone() for k, v in state_dict.items()} , os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) del state_dict gc.collect() if push_to_hub: if model_name is None: raise ValueError("""Please provide a `model_name` to push the model to the Hub.""" ) _a : Dict = AutoModelForCausalLM.from_pretrained(UpperCamelCase__ ) model.push_to_hub(UpperCamelCase__ , max_shard_size="""2GB""" ) tokenizer.push_to_hub(UpperCamelCase__ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '--repo_id', default=None, type=str, required=True, help='Repo ID from which to pull the checkpoint.' ) parser.add_argument( '--checkpoint_file', default=None, type=str, required=True, help='Name of the checkpoint file in the repo.' ) parser.add_argument( '--output_dir', default=None, type=str, required=True, help='Where to save the converted model.' ) parser.add_argument( '--tokenizer_file', default=None, type=str, help='Path to the tokenizer file to use (if not provided, only the model is converted).', ) parser.add_argument( '--size', default=None, type=str, help='Size of the model. Will be inferred from the `checkpoint_file` if not passed.', ) parser.add_argument( '--push_to_hub', action='store_true', help='Push to the Hub the converted model.', ) parser.add_argument( '--model_name', default=None, type=str, help='Name of the pushed model on the Hub, including the username / organization.', ) _snake_case = parser.parse_args() convert_rmkv_checkpoint_to_hf_format( args.repo_id, args.checkpoint_file, args.output_dir, size=args.size, tokenizer_file=args.tokenizer_file, push_to_hub=args.push_to_hub, model_name=args.model_name, )
294
1
"""simple docstring""" import numpy as np def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' return np.where(vector > 0 , UpperCamelCase__ , (alpha * (np.exp(UpperCamelCase__ ) - 1)) ) if __name__ == "__main__": import doctest doctest.testmod()
294
"""simple docstring""" import shutil import tempfile import unittest from transformers import ClapFeatureExtractor, ClapProcessor, RobertaTokenizer, RobertaTokenizerFast from transformers.testing_utils import require_sentencepiece, require_torchaudio from .test_feature_extraction_clap import floats_list @require_torchaudio @require_sentencepiece class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : int ) -> List[str]: _a : Any = """laion/clap-htsat-unfused""" _a : Union[str, Any] = tempfile.mkdtemp() def _lowercase ( self : List[Any] , **UpperCAmelCase__ : Any ) -> Dict: return RobertaTokenizer.from_pretrained(self.checkpoint , **UpperCAmelCase__ ) def _lowercase ( self : List[Any] , **UpperCAmelCase__ : List[str] ) -> int: return ClapFeatureExtractor.from_pretrained(self.checkpoint , **UpperCAmelCase__ ) def _lowercase ( self : List[Any] ) -> Tuple: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : List[str] ) -> Optional[int]: _a : List[str] = self.get_tokenizer() _a : Any = self.get_feature_extractor() _a : Optional[Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) _a : List[str] = ClapProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCAmelCase__ ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , UpperCAmelCase__ ) def _lowercase ( self : Dict ) -> Optional[int]: _a : Tuple = ClapProcessor(tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() ) processor.save_pretrained(self.tmpdirname ) _a : Dict = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) _a : Union[str, Any] = self.get_feature_extractor(do_normalize=UpperCAmelCase__ , padding_value=1.0 ) _a : Union[str, Any] = ClapProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=UpperCAmelCase__ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCAmelCase__ ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.feature_extractor , UpperCAmelCase__ ) def _lowercase ( self : List[str] ) -> Optional[Any]: _a : Optional[int] = self.get_feature_extractor() _a : Tuple = self.get_tokenizer() _a : List[Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) _a : Any = floats_list((3, 1000) ) _a : List[Any] = feature_extractor(UpperCAmelCase__ , return_tensors="""np""" ) _a : List[str] = processor(audios=UpperCAmelCase__ , return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def _lowercase ( self : Tuple ) -> Optional[int]: _a : List[str] = self.get_feature_extractor() _a : Any = self.get_tokenizer() _a : Any = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) _a : Optional[int] = """This is a test string""" _a : Tuple = processor(text=UpperCAmelCase__ ) _a : int = tokenizer(UpperCAmelCase__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _lowercase ( self : List[Any] ) -> Any: _a : str = self.get_feature_extractor() _a : List[str] = self.get_tokenizer() _a : List[Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) _a : Any = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _a : Dict = processor.batch_decode(UpperCAmelCase__ ) _a : Any = tokenizer.batch_decode(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def _lowercase ( self : Dict ) -> List[str]: _a : str = self.get_feature_extractor() _a : Optional[Any] = self.get_tokenizer() _a : Union[str, Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) self.assertListEqual( processor.model_input_names[2:] , feature_extractor.model_input_names , msg="""`processor` and `feature_extractor` model input names do not match""" , )
294
1
"""simple docstring""" import copy import inspect import unittest from transformers import AutoBackbone from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import require_timm, require_torch, torch_device from transformers.utils.import_utils import is_torch_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor if is_torch_available(): import torch from transformers import TimmBackbone, TimmBackboneConfig from ...test_pipeline_mixin import PipelineTesterMixin class UpperCamelCase : def __init__( self : List[Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : int=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Tuple=None , UpperCAmelCase__ : Any="resnet50" , UpperCAmelCase__ : Optional[int]=3 , UpperCAmelCase__ : List[str]=32 , UpperCAmelCase__ : Dict=3 , UpperCAmelCase__ : Optional[Any]=True , UpperCAmelCase__ : int=True , ) -> Any: _a : Union[str, Any] = parent _a : Tuple = out_indices if out_indices is not None else [4] _a : List[str] = stage_names _a : List[Any] = out_features _a : Dict = backbone _a : Dict = batch_size _a : Any = image_size _a : Tuple = num_channels _a : Optional[int] = use_pretrained_backbone _a : Dict = is_training def _lowercase ( self : Union[str, Any] ) -> Optional[Any]: _a : Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _a : int = self.get_config() return config, pixel_values def _lowercase ( self : List[Any] ) -> Optional[Any]: return TimmBackboneConfig( image_size=self.image_size , num_channels=self.num_channels , out_features=self.out_features , out_indices=self.out_indices , stage_names=self.stage_names , use_pretrained_backbone=self.use_pretrained_backbone , backbone=self.backbone , ) def _lowercase ( self : Tuple , UpperCAmelCase__ : int , UpperCAmelCase__ : Optional[int] ) -> List[str]: _a : List[str] = TimmBackbone(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() with torch.no_grad(): _a : str = model(UpperCAmelCase__ ) self.parent.assertEqual( result.feature_map[-1].shape , (self.batch_size, model.channels[-1], 14, 14) , ) def _lowercase ( self : List[Any] ) -> Optional[int]: _a : Dict = self.prepare_config_and_inputs() _a , _a : Tuple = config_and_inputs _a : Any = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch @require_timm class UpperCamelCase ( snake_case_ , snake_case_ , snake_case_ , unittest.TestCase ): UpperCamelCase : Any = (TimmBackbone,) if is_torch_available() else () UpperCamelCase : int = {'''feature-extraction''': TimmBackbone} if is_torch_available() else {} UpperCamelCase : List[Any] = False UpperCamelCase : Tuple = False UpperCamelCase : Union[str, Any] = False UpperCamelCase : Tuple = False def _lowercase ( self : Optional[Any] ) -> List[Any]: _a : str = TimmBackboneModelTester(self ) _a : Optional[int] = ConfigTester(self , config_class=UpperCAmelCase__ , has_text_modality=UpperCAmelCase__ ) def _lowercase ( self : List[str] ) -> Optional[Any]: self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _lowercase ( self : Any ) -> Tuple: _a : Optional[Any] = """resnet18""" _a : Union[str, Any] = """microsoft/resnet-18""" _a : int = AutoBackbone.from_pretrained(UpperCAmelCase__ , use_timm_backbone=UpperCAmelCase__ ) _a : Union[str, Any] = AutoBackbone.from_pretrained(UpperCAmelCase__ ) self.assertEqual(len(timm_model.out_features ) , len(transformers_model.out_features ) ) self.assertEqual(len(timm_model.stage_names ) , len(transformers_model.stage_names ) ) self.assertEqual(timm_model.channels , transformers_model.channels ) # Out indices are set to the last layer by default. For timm models, we don't know # the number of layers in advance, so we set it to (-1,), whereas for transformers # models, we set it to [len(stage_names) - 1] (kept for backward compatibility). self.assertEqual(timm_model.out_indices , (-1,) ) self.assertEqual(transformers_model.out_indices , [len(timm_model.stage_names ) - 1] ) _a : int = AutoBackbone.from_pretrained(UpperCAmelCase__ , use_timm_backbone=UpperCAmelCase__ , out_indices=[1, 2, 3] ) _a : Optional[Any] = AutoBackbone.from_pretrained(UpperCAmelCase__ , out_indices=[1, 2, 3] ) self.assertEqual(timm_model.out_indices , transformers_model.out_indices ) self.assertEqual(len(timm_model.out_features ) , len(transformers_model.out_features ) ) self.assertEqual(timm_model.channels , transformers_model.channels ) @unittest.skip("""TimmBackbone doesn't support feed forward chunking""" ) def _lowercase ( self : List[Any] ) -> int: pass @unittest.skip("""TimmBackbone doesn't have num_hidden_layers attribute""" ) def _lowercase ( self : Optional[int] ) -> int: pass @unittest.skip("""TimmBackbone initialization is managed on the timm side""" ) def _lowercase ( self : Optional[int] ) -> List[Any]: pass @unittest.skip("""TimmBackbone models doesn't have inputs_embeds""" ) def _lowercase ( self : Optional[int] ) -> Any: pass @unittest.skip("""TimmBackbone models doesn't have inputs_embeds""" ) def _lowercase ( self : Tuple ) -> Tuple: pass @unittest.skip("""TimmBackbone model cannot be created without specifying a backbone checkpoint""" ) def _lowercase ( self : Union[str, Any] ) -> Any: pass @unittest.skip("""Only checkpoints on timm can be loaded into TimmBackbone""" ) def _lowercase ( self : Dict ) -> str: pass @unittest.skip("""model weights aren't tied in TimmBackbone.""" ) def _lowercase ( self : Dict ) -> Tuple: pass @unittest.skip("""model weights aren't tied in TimmBackbone.""" ) def _lowercase ( self : Dict ) -> Tuple: pass @unittest.skip("""Only checkpoints on timm can be loaded into TimmBackbone""" ) def _lowercase ( self : Union[str, Any] ) -> Optional[Any]: pass @unittest.skip("""Only checkpoints on timm can be loaded into TimmBackbone""" ) def _lowercase ( self : str ) -> List[str]: pass @unittest.skip("""TimmBackbone doesn't have hidden size info in its configuration.""" ) def _lowercase ( self : List[Any] ) -> Dict: pass @unittest.skip("""TimmBackbone doesn't support output_attentions.""" ) def _lowercase ( self : Union[str, Any] ) -> Tuple: pass @unittest.skip("""Safetensors is not supported by timm.""" ) def _lowercase ( self : List[Any] ) -> Dict: pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def _lowercase ( self : Tuple ) -> Dict: pass def _lowercase ( self : List[str] ) -> Optional[int]: _a , _a : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Union[str, Any] = model_class(UpperCAmelCase__ ) _a : Tuple = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _a : Optional[int] = [*signature.parameters.keys()] _a : str = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCAmelCase__ ) def _lowercase ( self : int ) -> Dict: _a , _a : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() _a : str = True _a : Union[str, Any] = self.has_attentions # no need to test all models as different heads yield the same functionality _a : Optional[Any] = self.all_model_classes[0] _a : Tuple = model_class(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) _a : List[str] = self._prepare_for_class(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Optional[int] = model(**UpperCAmelCase__ ) _a : Any = outputs[0][-1] # Encoder-/Decoder-only models _a : str = outputs.hidden_states[0] hidden_states.retain_grad() if self.has_attentions: _a : List[str] = outputs.attentions[0] attentions.retain_grad() output.flatten()[0].backward(retain_graph=UpperCAmelCase__ ) self.assertIsNotNone(hidden_states.grad ) if self.has_attentions: self.assertIsNotNone(attentions.grad ) def _lowercase ( self : str ) -> Tuple: _a , _a : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : int = model_class(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() _a : Optional[int] = model(**UpperCAmelCase__ ) self.assertEqual(len(result.feature_maps ) , len(config.out_indices ) ) self.assertEqual(len(model.channels ) , len(config.out_indices ) ) # Check output of last stage is taken if out_features=None, out_indices=None _a : Dict = copy.deepcopy(UpperCAmelCase__ ) _a : Dict = None _a : Optional[Any] = model_class(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() _a : List[Any] = model(**UpperCAmelCase__ ) self.assertEqual(len(result.feature_maps ) , 1 ) self.assertEqual(len(model.channels ) , 1 ) # Check backbone can be initialized with fresh weights _a : Any = copy.deepcopy(UpperCAmelCase__ ) _a : Tuple = False _a : Optional[Any] = model_class(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() _a : Any = model(**UpperCAmelCase__ )
294
"""simple docstring""" import warnings from ...utils import logging from .image_processing_clip import CLIPImageProcessor _snake_case = logging.get_logger(__name__) class UpperCamelCase ( snake_case_ ): def __init__( self : Any , *UpperCAmelCase__ : Optional[Any] , **UpperCAmelCase__ : Tuple ) -> None: warnings.warn( """The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use CLIPImageProcessor instead.""" , UpperCAmelCase__ , ) super().__init__(*UpperCAmelCase__ , **UpperCAmelCase__ )
294
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) _snake_case = { 'configuration_trocr': ['TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TrOCRConfig'], 'processing_trocr': ['TrOCRProcessor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'TROCR_PRETRAINED_MODEL_ARCHIVE_LIST', 'TrOCRForCausalLM', 'TrOCRPreTrainedModel', ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys _snake_case = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
294
"""simple docstring""" import unittest import numpy as np def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = None , ): '''simple docstring''' _a : List[Any] = np.shape(UpperCamelCase__ ) _a : Any = np.shape(UpperCamelCase__ ) _a : Union[str, Any] = np.shape(UpperCamelCase__ ) if shape_a[0] != shape_b[0]: _a : int = ( """Expected the same number of rows for A and B. """ F"""Instead found A of size {shape_a} and B of size {shape_b}""" ) raise ValueError(UpperCamelCase__ ) if shape_b[1] != shape_c[1]: _a : Tuple = ( """Expected the same number of columns for B and C. """ F"""Instead found B of size {shape_b} and C of size {shape_c}""" ) raise ValueError(UpperCamelCase__ ) _a : int = pseudo_inv if a_inv is None: try: _a : Optional[int] = np.linalg.inv(UpperCamelCase__ ) except np.linalg.LinAlgError: raise ValueError( """Input matrix A is not invertible. Cannot compute Schur complement.""" ) return mat_c - mat_b.T @ a_inv @ mat_b class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : int ) -> None: _a : str = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) _a : Tuple = np.array([[0, 3], [3, 0], [2, 3]] ) _a : Optional[int] = np.array([[2, 1], [6, 3]] ) _a : Optional[Any] = schur_complement(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) _a : Union[str, Any] = np.block([[a, b], [b.T, c]] ) _a : int = np.linalg.det(UpperCAmelCase__ ) _a : Union[str, Any] = np.linalg.det(UpperCAmelCase__ ) _a : List[Any] = np.linalg.det(UpperCAmelCase__ ) self.assertAlmostEqual(UpperCAmelCase__ , det_a * det_s ) def _lowercase ( self : int ) -> None: _a : Optional[int] = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) _a : Optional[int] = np.array([[0, 3], [3, 0], [2, 3]] ) _a : Union[str, Any] = np.array([[2, 1], [6, 3]] ) with self.assertRaises(UpperCAmelCase__ ): schur_complement(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) def _lowercase ( self : List[Any] ) -> None: _a : Any = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) _a : Dict = np.array([[0, 3], [3, 0], [2, 3]] ) _a : List[Any] = np.array([[2, 1, 3], [6, 3, 5]] ) with self.assertRaises(UpperCAmelCase__ ): schur_complement(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) if __name__ == "__main__": import doctest doctest.testmod() unittest.main()
294
1
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() _snake_case = logging.get_logger(__name__) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__=False ): '''simple docstring''' _a : int = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""blocks.{i}.norm1.weight""", F"""deit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""blocks.{i}.norm1.bias""", F"""deit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((F"""blocks.{i}.attn.proj.weight""", F"""deit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.attn.proj.bias""", F"""deit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((F"""blocks.{i}.norm2.weight""", F"""deit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""blocks.{i}.norm2.bias""", F"""deit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.weight""", F"""deit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.bias""", F"""deit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.weight""", F"""deit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.bias""", F"""deit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ("""cls_token""", """deit.embeddings.cls_token"""), ("""dist_token""", """deit.embeddings.distillation_token"""), ("""patch_embed.proj.weight""", """deit.embeddings.patch_embeddings.projection.weight"""), ("""patch_embed.proj.bias""", """deit.embeddings.patch_embeddings.projection.bias"""), ("""pos_embed""", """deit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""norm.weight""", """layernorm.weight"""), ("""norm.bias""", """layernorm.bias"""), ("""pre_logits.fc.weight""", """pooler.dense.weight"""), ("""pre_logits.fc.bias""", """pooler.dense.bias"""), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" _a : List[Any] = [(pair[0], pair[1][4:]) if pair[1].startswith("""deit""" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("""norm.weight""", """deit.layernorm.weight"""), ("""norm.bias""", """deit.layernorm.bias"""), ("""head.weight""", """cls_classifier.weight"""), ("""head.bias""", """cls_classifier.bias"""), ("""head_dist.weight""", """distillation_classifier.weight"""), ("""head_dist.bias""", """distillation_classifier.bias"""), ] ) return rename_keys def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=False ): '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: _a : Optional[Any] = """""" else: _a : Any = """deit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _a : Tuple = state_dict.pop(F"""blocks.{i}.attn.qkv.weight""" ) _a : str = state_dict.pop(F"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict _a : Dict = in_proj_weight[ : config.hidden_size, : ] _a : List[Any] = in_proj_bias[: config.hidden_size] _a : int = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _a : Tuple = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _a : Tuple = in_proj_weight[ -config.hidden_size :, : ] _a : Any = in_proj_bias[-config.hidden_size :] def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : int = dct.pop(UpperCamelCase__ ) _a : int = val def lowerCAmelCase__ ( ): '''simple docstring''' _a : Union[str, Any] = """http://images.cocodataset.org/val2017/000000039769.jpg""" _a : Union[str, Any] = Image.open(requests.get(UpperCamelCase__ , stream=UpperCamelCase__ ).raw ) return im @torch.no_grad() def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Union[str, Any] = DeiTConfig() # all deit models have fine-tuned heads _a : Optional[int] = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size _a : Tuple = 1_0_0_0 _a : Dict = """huggingface/label-files""" _a : Union[str, Any] = """imagenet-1k-id2label.json""" _a : Any = json.load(open(hf_hub_download(UpperCamelCase__ , UpperCamelCase__ , repo_type="""dataset""" ) , """r""" ) ) _a : Any = {int(UpperCamelCase__ ): v for k, v in idalabel.items()} _a : Tuple = idalabel _a : List[Any] = {v: k for k, v in idalabel.items()} _a : Union[str, Any] = int(deit_name[-6:-4] ) _a : List[str] = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("""tiny""" ): _a : Dict = 1_9_2 _a : Dict = 7_6_8 _a : Dict = 1_2 _a : List[str] = 3 elif deit_name[9:].startswith("""small""" ): _a : Optional[Any] = 3_8_4 _a : Dict = 1_5_3_6 _a : List[str] = 1_2 _a : List[Any] = 6 if deit_name[9:].startswith("""base""" ): pass elif deit_name[4:].startswith("""large""" ): _a : Optional[Any] = 1_0_2_4 _a : List[str] = 4_0_9_6 _a : Optional[Any] = 2_4 _a : Union[str, Any] = 1_6 # load original model from timm _a : Optional[int] = timm.create_model(UpperCamelCase__ , pretrained=UpperCamelCase__ ) timm_model.eval() # load state_dict of original model, remove and rename some keys _a : Optional[int] = timm_model.state_dict() _a : str = create_rename_keys(UpperCamelCase__ , UpperCamelCase__ ) for src, dest in rename_keys: rename_key(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) read_in_q_k_v(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # load HuggingFace model _a : List[Any] = DeiTForImageClassificationWithTeacher(UpperCamelCase__ ).eval() model.load_state_dict(UpperCamelCase__ ) # Check outputs on an image, prepared by DeiTImageProcessor _a : List[Any] = int( (2_5_6 / 2_2_4) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 _a : Any = DeiTImageProcessor(size=UpperCamelCase__ , crop_size=config.image_size ) _a : List[str] = image_processor(images=prepare_img() , return_tensors="""pt""" ) _a : List[str] = encoding["""pixel_values"""] _a : str = model(UpperCamelCase__ ) _a : int = timm_model(UpperCamelCase__ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(UpperCamelCase__ , outputs.logits , atol=1e-3 ) Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) print(F"""Saving model {deit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(UpperCamelCase__ ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(UpperCamelCase__ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '--deit_name', default='vit_deit_base_distilled_patch16_224', type=str, help='Name of the DeiT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) _snake_case = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
294
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _snake_case = logging.get_logger(__name__) _snake_case = torch.device('cpu') def lowerCAmelCase__ ( ): '''simple docstring''' _a : int = """http://images.cocodataset.org/val2017/000000039769.jpg""" _a : Dict = Image.open(requests.get(UpperCamelCase__ , stream=UpperCamelCase__ ).raw ) return im def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' if swiftformer_name == "swiftformer_xs": return torch.tensor([-2.17_03e00, 2.11_07e00, -2.08_11e00, 8.86_85e-01, 2.43_60e-01] ) elif swiftformer_name == "swiftformer_s": return torch.tensor([3.96_36e-01, 2.34_78e-01, -1.69_63e00, -1.73_81e00, -8.63_37e-01] ) elif swiftformer_name == "swiftformer_l1": return torch.tensor([-4.27_68e-01, -4.74_29e-01, -1.08_97e00, -1.02_48e00, 3.55_23e-02] ) elif swiftformer_name == "swiftformer_l3": return torch.tensor([-2.53_30e-01, 2.42_11e-01, -6.01_85e-01, -8.27_89e-01, -6.04_46e-02] ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Any = dct.pop(UpperCamelCase__ ) _a : Dict = val def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : Tuple = [] for k in state_dict.keys(): _a : Any = k if ".pwconv" in k: _a : int = k_new.replace(""".pwconv""" , """.point_wise_conv""" ) if ".dwconv" in k: _a : List[str] = k_new.replace(""".dwconv""" , """.depth_wise_conv""" ) if ".Proj." in k: _a : Optional[int] = k_new.replace(""".Proj.""" , """.proj.""" ) if "patch_embed" in k_new: _a : Tuple = k_new.replace("""patch_embed""" , """swiftformer.patch_embed.patch_embedding""" ) if "network" in k_new: _a : int = k_new.split(""".""" ) if ls[2].isdigit(): _a : Union[str, Any] = """swiftformer.encoder.network.""" + ls[1] + """.blocks.""" + ls[2] + """.""" + """.""".join(ls[3:] ) else: _a : Tuple = k_new.replace("""network""" , """swiftformer.encoder.network""" ) rename_keys.append((k, k_new) ) return rename_keys @torch.no_grad() def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Tuple = SwiftFormerConfig() # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size _a : Optional[int] = 1_0_0_0 _a : Optional[Any] = """huggingface/label-files""" _a : Optional[Any] = """imagenet-1k-id2label.json""" _a : List[str] = json.load(open(hf_hub_download(UpperCamelCase__ , UpperCamelCase__ , repo_type="""dataset""" ) , """r""" ) ) _a : Optional[Any] = {int(UpperCamelCase__ ): v for k, v in idalabel.items()} _a : Dict = idalabel _a : Optional[int] = {v: k for k, v in idalabel.items()} # size of the architecture if swiftformer_name == "swiftformer_xs": _a : Any = [3, 3, 6, 4] _a : int = [4_8, 5_6, 1_1_2, 2_2_0] elif swiftformer_name == "swiftformer_s": _a : Any = [3, 3, 9, 6] _a : List[str] = [4_8, 6_4, 1_6_8, 2_2_4] elif swiftformer_name == "swiftformer_l1": _a : List[Any] = [4, 3, 1_0, 5] _a : Optional[int] = [4_8, 9_6, 1_9_2, 3_8_4] elif swiftformer_name == "swiftformer_l3": _a : List[Any] = [4, 4, 1_2, 6] _a : Optional[Any] = [6_4, 1_2_8, 3_2_0, 5_1_2] # load state_dict of original model, remove and rename some keys if original_ckpt: if original_ckpt.startswith("""https""" ): _a : Tuple = torch.hub.load_state_dict_from_url(UpperCamelCase__ , map_location="""cpu""" , check_hash=UpperCamelCase__ ) else: _a : Dict = torch.load(UpperCamelCase__ , map_location="""cpu""" ) _a : int = checkpoint _a : Optional[Any] = create_rename_keys(UpperCamelCase__ ) for rename_key_src, rename_key_dest in rename_keys: rename_key(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # load HuggingFace model _a : Any = SwiftFormerForImageClassification(UpperCamelCase__ ).eval() hf_model.load_state_dict(UpperCamelCase__ ) # prepare test inputs _a : Any = prepare_img() _a : Union[str, Any] = ViTImageProcessor.from_pretrained("""preprocessor_config""" ) _a : Optional[int] = processor(images=UpperCamelCase__ , return_tensors="""pt""" ) # compare outputs from both models _a : Dict = get_expected_output(UpperCamelCase__ ) _a : int = hf_model(inputs["""pixel_values"""] ).logits assert hf_logits.shape == torch.Size([1, 1_0_0_0] ) assert torch.allclose(hf_logits[0, 0:5] , UpperCamelCase__ , atol=1e-3 ) Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) print(F"""Saving model {swiftformer_name} to {pytorch_dump_folder_path}""" ) hf_model.save_pretrained(UpperCamelCase__ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '--swiftformer_name', default='swiftformer_xs', choices=['swiftformer_xs', 'swiftformer_s', 'swiftformer_l1', 'swiftformer_l3'], type=str, help='Name of the SwiftFormer model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default='./converted_outputs/', type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument('--original_ckpt', default=None, type=str, help='Path to the original model checkpoint.') _snake_case = parser.parse_args() convert_swiftformer_checkpoint(args.swiftformer_name, args.pytorch_dump_folder_path, args.original_ckpt)
294
1
"""simple docstring""" import jax.numpy as jnp from ...utils import logging from ..ta.modeling_flax_ta import FlaxTaEncoderModel, FlaxTaForConditionalGeneration, FlaxTaModel from .configuration_mta import MTaConfig _snake_case = logging.get_logger(__name__) _snake_case = 'T5Config' def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Any = jnp.zeros_like(UpperCamelCase__ ) _a : Tuple = shifted_input_ids.at[:, 1:].set(input_ids[:, :-1] ) _a : Union[str, Any] = shifted_input_ids.at[:, 0].set(UpperCamelCase__ ) _a : Union[str, Any] = jnp.where(shifted_input_ids == -1_0_0 , UpperCamelCase__ , UpperCamelCase__ ) return shifted_input_ids class UpperCamelCase ( snake_case_ ): UpperCamelCase : List[Any] = '''mt5''' UpperCamelCase : Optional[int] = MTaConfig class UpperCamelCase ( snake_case_ ): UpperCamelCase : Dict = '''mt5''' UpperCamelCase : Union[str, Any] = MTaConfig class UpperCamelCase ( snake_case_ ): UpperCamelCase : int = '''mt5''' UpperCamelCase : List[Any] = MTaConfig
294
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case = { 'configuration_perceiver': ['PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PerceiverConfig', 'PerceiverOnnxConfig'], 'tokenization_perceiver': ['PerceiverTokenizer'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ['PerceiverFeatureExtractor'] _snake_case = ['PerceiverImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST', 'PerceiverForImageClassificationConvProcessing', 'PerceiverForImageClassificationFourier', 'PerceiverForImageClassificationLearned', 'PerceiverForMaskedLM', 'PerceiverForMultimodalAutoencoding', 'PerceiverForOpticalFlow', 'PerceiverForSequenceClassification', 'PerceiverLayer', 'PerceiverModel', 'PerceiverPreTrainedModel', ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
294
1
"""simple docstring""" import os import shutil import tempfile import unittest import numpy as np from transformers import AutoTokenizer, BarkProcessor from transformers.testing_utils import require_torch, slow @require_torch class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : Tuple ) -> List[Any]: _a : Dict = """ylacombe/bark-small""" _a : Union[str, Any] = tempfile.mkdtemp() _a : Dict = """en_speaker_1""" _a : Any = """This is a test string""" _a : str = """speaker_embeddings_path.json""" _a : Union[str, Any] = """speaker_embeddings""" def _lowercase ( self : Any , **UpperCAmelCase__ : Union[str, Any] ) -> Optional[Any]: return AutoTokenizer.from_pretrained(self.checkpoint , **UpperCAmelCase__ ) def _lowercase ( self : Dict ) -> Any: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Optional[Any] ) -> Dict: _a : List[Any] = self.get_tokenizer() _a : List[Any] = BarkProcessor(tokenizer=UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) _a : Any = BarkProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) @slow def _lowercase ( self : str ) -> Any: _a : Dict = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) processor.save_pretrained( self.tmpdirname , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , speaker_embeddings_directory=self.speaker_embeddings_directory , ) _a : Dict = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) _a : str = BarkProcessor.from_pretrained( self.tmpdirname , self.speaker_embeddings_dict_path , bos_token="""(BOS)""" , eos_token="""(EOS)""" , ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) def _lowercase ( self : Tuple ) -> Any: _a : Union[str, Any] = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) _a : Dict = 35 _a : Any = 2 _a : str = 8 _a : Tuple = { """semantic_prompt""": np.ones(UpperCAmelCase__ ), """coarse_prompt""": np.ones((nb_codebooks_coarse, seq_len) ), """fine_prompt""": np.ones((nb_codebooks_total, seq_len) ), } # test providing already loaded voice_preset _a : Optional[int] = processor(text=self.input_string , voice_preset=UpperCAmelCase__ ) _a : Any = inputs["""history_prompt"""] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(UpperCAmelCase__ , np.array([] ) ).tolist() ) # test loading voice preset from npz file _a : str = os.path.join(self.tmpdirname , """file.npz""" ) np.savez(UpperCAmelCase__ , **UpperCAmelCase__ ) _a : str = processor(text=self.input_string , voice_preset=UpperCAmelCase__ ) _a : List[Any] = inputs["""history_prompt"""] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(UpperCAmelCase__ , np.array([] ) ).tolist() ) # test loading voice preset from the hub _a : Optional[Any] = processor(text=self.input_string , voice_preset=self.voice_preset ) def _lowercase ( self : int ) -> List[str]: _a : Union[str, Any] = self.get_tokenizer() _a : Union[str, Any] = BarkProcessor(tokenizer=UpperCAmelCase__ ) _a : Any = processor(text=self.input_string ) _a : List[str] = tokenizer( self.input_string , padding="""max_length""" , max_length=256 , add_special_tokens=UpperCAmelCase__ , return_attention_mask=UpperCAmelCase__ , return_token_type_ids=UpperCAmelCase__ , ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key].squeeze().tolist() )
294
"""simple docstring""" import cva import numpy as np class UpperCamelCase : def __init__( self : Optional[int] , UpperCAmelCase__ : float , UpperCAmelCase__ : int ) -> Dict: if k in (0.0_4, 0.0_6): _a : List[str] = k _a : List[Any] = window_size else: raise ValueError("""invalid k value""" ) def __str__( self : Dict ) -> str: return str(self.k ) def _lowercase ( self : int , UpperCAmelCase__ : str ) -> tuple[cva.Mat, list[list[int]]]: _a : Dict = cva.imread(UpperCAmelCase__ , 0 ) _a , _a : List[Any] = img.shape _a : list[list[int]] = [] _a : List[Any] = img.copy() _a : int = cva.cvtColor(UpperCAmelCase__ , cva.COLOR_GRAY2RGB ) _a , _a : Any = np.gradient(UpperCAmelCase__ ) _a : Tuple = dx**2 _a : Union[str, Any] = dy**2 _a : Union[str, Any] = dx * dy _a : int = 0.0_4 _a : List[str] = self.window_size // 2 for y in range(UpperCAmelCase__ , h - offset ): for x in range(UpperCAmelCase__ , w - offset ): _a : str = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _a : List[Any] = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _a : Tuple = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _a : Any = (wxx * wyy) - (wxy**2) _a : Tuple = wxx + wyy _a : Any = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 255 ) return color_img, corner_list if __name__ == "__main__": _snake_case = HarrisCorner(0.04, 3) _snake_case , _snake_case = edge_detect.detect('path_to_image') cva.imwrite('detect.png', color_img)
294
1
"""simple docstring""" def lowerCAmelCase__ ( UpperCamelCase__ = 1_0_0_0_0_0_0 ): '''simple docstring''' _a : Optional[int] = set(range(3 , UpperCamelCase__ , 2 ) ) primes.add(2 ) for p in range(3 , UpperCamelCase__ , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , UpperCamelCase__ , UpperCamelCase__ ) ) ) _a : List[Any] = [float(UpperCamelCase__ ) for n in range(limit + 1 )] for p in primes: for n in range(UpperCamelCase__ , limit + 1 , UpperCamelCase__ ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(F'''{solution() = }''')
294
"""simple docstring""" def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' # Check if the input is valid if not len(UpperCamelCase__ ) == len(UpperCamelCase__ ) == 3: raise ValueError("""Please enter a valid equation.""" ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError("""Both a & b of two equations can't be zero.""" ) # Extract the coefficients _a , _a , _a : Any = equationa _a , _a , _a : Tuple = equationa # Calculate the determinants of the matrices _a : int = aa * ba - aa * ba _a : str = ca * ba - ca * ba _a : str = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError("""Infinite solutions. (Consistent system)""" ) else: raise ValueError("""No solution. (Inconsistent system)""" ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: _a : Dict = determinant_x / determinant _a : str = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
294
1
"""simple docstring""" import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, StableDiffusionAttendAndExcitePipeline, UNetaDConditionModel, ) from diffusers.utils import load_numpy, skip_mps, slow from diffusers.utils.testing_utils import require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin _snake_case = False @skip_mps class UpperCamelCase ( snake_case_ , snake_case_ , snake_case_ , unittest.TestCase ): UpperCamelCase : List[Any] = StableDiffusionAttendAndExcitePipeline UpperCamelCase : Union[str, Any] = False UpperCamelCase : Tuple = TEXT_TO_IMAGE_PARAMS UpperCamelCase : int = TEXT_TO_IMAGE_BATCH_PARAMS.union({'''token_indices'''} ) UpperCamelCase : List[str] = TEXT_TO_IMAGE_IMAGE_PARAMS UpperCamelCase : List[Any] = TEXT_TO_IMAGE_IMAGE_PARAMS @classmethod def _lowercase ( cls : Union[str, Any] ) -> Optional[Any]: super().setUpClass() torch.use_deterministic_algorithms(UpperCAmelCase__ ) @classmethod def _lowercase ( cls : List[Any] ) -> Tuple: super().tearDownClass() torch.use_deterministic_algorithms(UpperCAmelCase__ ) def _lowercase ( self : List[str] ) -> Optional[int]: torch.manual_seed(0 ) _a : Optional[Any] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=1 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=UpperCAmelCase__ , ) _a : List[str] = DDIMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule="""scaled_linear""" , clip_sample=UpperCAmelCase__ , set_alpha_to_one=UpperCAmelCase__ , ) torch.manual_seed(0 ) _a : str = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) _a : Optional[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act="""gelu""" , projection_dim=512 , ) _a : int = CLIPTextModel(UpperCAmelCase__ ) _a : Union[str, Any] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _a : Optional[Any] = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def _lowercase ( self : Union[str, Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : Optional[int]=0 ) -> Any: if str(UpperCAmelCase__ ).startswith("""mps""" ): _a : Dict = torch.manual_seed(UpperCAmelCase__ ) else: _a : str = torch.Generator(device=UpperCAmelCase__ ).manual_seed(UpperCAmelCase__ ) _a : Any = { """prompt""": """a cat and a frog""", """token_indices""": [2, 5], """generator""": generator, """num_inference_steps""": 1, """guidance_scale""": 6.0, """output_type""": """numpy""", """max_iter_to_alter""": 2, """thresholds""": {0: 0.7}, } return inputs def _lowercase ( self : Dict ) -> Optional[Any]: _a : Optional[int] = """cpu""" _a : int = self.get_dummy_components() _a : Optional[Any] = self.pipeline_class(**UpperCAmelCase__ ) pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : List[str] = self.get_dummy_inputs(UpperCAmelCase__ ) _a : Optional[int] = pipe(**UpperCAmelCase__ ).images _a : List[Any] = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 64, 64, 3) ) _a : str = np.array( [0.6_3_9_0_5_3_6_4, 0.6_2_8_9_7_3_0_7, 0.4_8_5_9_9_0_1_7, 0.5_1_3_3_6_2_4, 0.5_5_5_0_0_4_8, 0.4_5_7_6_9_5_1_6, 0.5_0_3_2_6_9_7_3, 0.5_0_2_3_1_3_9, 0.4_5_3_8_4_4_9_6] ) _a : Any = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(UpperCAmelCase__ , 1E-3 ) def _lowercase ( self : int ) -> int: super().test_cpu_offload_forward_pass(expected_max_diff=5E-4 ) def _lowercase ( self : List[str] ) -> Optional[Any]: # NOTE: Larger batch sizes cause this test to timeout, only test on smaller batches self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def _lowercase ( self : List[str] ) -> List[str]: self._test_inference_batch_single_identical(batch_size=2 , expected_max_diff=7E-4 ) def _lowercase ( self : str ) -> Any: super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) def _lowercase ( self : Optional[int] ) -> str: super().test_pt_np_pil_outputs_equivalent(expected_max_diff=5E-4 ) def _lowercase ( self : Optional[Any] ) -> Any: super().test_save_load_local(expected_max_difference=5E-4 ) def _lowercase ( self : Any ) -> Optional[Any]: super().test_save_load_optional_components(expected_max_difference=4E-4 ) @require_torch_gpu @slow class UpperCamelCase ( unittest.TestCase ): @classmethod def _lowercase ( cls : int ) -> List[Any]: super().setUpClass() torch.use_deterministic_algorithms(UpperCAmelCase__ ) @classmethod def _lowercase ( cls : str ) -> int: super().tearDownClass() torch.use_deterministic_algorithms(UpperCAmelCase__ ) def _lowercase ( self : Union[str, Any] ) -> Union[str, Any]: super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self : int ) -> Union[str, Any]: _a : Dict = torch.manual_seed(51 ) _a : Union[str, Any] = StableDiffusionAttendAndExcitePipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , safety_checker=UpperCAmelCase__ , torch_dtype=torch.floataa ) pipe.to("""cuda""" ) _a : List[Any] = """a painting of an elephant with glasses""" _a : int = [5, 7] _a : Optional[Any] = pipe( prompt=UpperCAmelCase__ , token_indices=UpperCAmelCase__ , guidance_scale=7.5 , generator=UpperCAmelCase__ , num_inference_steps=5 , max_iter_to_alter=5 , output_type="""numpy""" , ).images[0] _a : int = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/attend-and-excite/elephant_glasses.npy""" ) assert np.abs((expected_image - image).max() ) < 5E-1
294
"""simple docstring""" _snake_case = '\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n' _snake_case = [{'type': 'code', 'content': INSTALL_CONTENT}] _snake_case = { '{processor_class}': 'FakeProcessorClass', '{model_class}': 'FakeModelClass', '{object_class}': 'FakeObjectClass', }
294
1
"""simple docstring""" import inspect import re from hashlib import shaaaa from typing import Dict, List from .arrow import arrow from .audiofolder import audiofolder from .csv import csv from .imagefolder import imagefolder from .json import json from .pandas import pandas from .parquet import parquet from .sql import sql # noqa F401 from .text import text def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : Tuple = [] for line in lines: _a : Optional[int] = re.sub(R"""#.*""" , """""" , UpperCamelCase__ ) # remove comments if line: filtered_lines.append(UpperCamelCase__ ) _a : Tuple = """\n""".join(UpperCamelCase__ ) # Make a hash from all this code _a : str = full_str.encode("""utf-8""" ) return shaaaa(UpperCamelCase__ ).hexdigest() # get importable module names and hash for caching _snake_case = { 'csv': (csv.__name__, _hash_python_lines(inspect.getsource(csv).splitlines())), 'json': (json.__name__, _hash_python_lines(inspect.getsource(json).splitlines())), 'pandas': (pandas.__name__, _hash_python_lines(inspect.getsource(pandas).splitlines())), 'parquet': (parquet.__name__, _hash_python_lines(inspect.getsource(parquet).splitlines())), 'arrow': (arrow.__name__, _hash_python_lines(inspect.getsource(arrow).splitlines())), 'text': (text.__name__, _hash_python_lines(inspect.getsource(text).splitlines())), 'imagefolder': (imagefolder.__name__, _hash_python_lines(inspect.getsource(imagefolder).splitlines())), 'audiofolder': (audiofolder.__name__, _hash_python_lines(inspect.getsource(audiofolder).splitlines())), } # Used to infer the module to use based on the data files extensions _snake_case = { '.csv': ('csv', {}), '.tsv': ('csv', {'sep': '\t'}), '.json': ('json', {}), '.jsonl': ('json', {}), '.parquet': ('parquet', {}), '.arrow': ('arrow', {}), '.txt': ('text', {}), } _EXTENSION_TO_MODULE.update({ext: ('imagefolder', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('imagefolder', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext: ('audiofolder', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('audiofolder', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _snake_case = {'imagefolder', 'audiofolder'} # Used to filter data files based on extensions given a module name _snake_case = {} for _ext, (_module, _) in _EXTENSION_TO_MODULE.items(): _MODULE_TO_EXTENSIONS.setdefault(_module, []).append(_ext) _MODULE_TO_EXTENSIONS["imagefolder"].append('.zip') _MODULE_TO_EXTENSIONS["audiofolder"].append('.zip')
294
"""simple docstring""" import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json', } class UpperCamelCase ( snake_case_ ): UpperCamelCase : int = '''mvp''' UpperCamelCase : Union[str, Any] = ['''past_key_values'''] UpperCamelCase : Any = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : List[str] , UpperCAmelCase__ : List[str]=50267 , UpperCAmelCase__ : Optional[Any]=1024 , UpperCAmelCase__ : Tuple=12 , UpperCAmelCase__ : Optional[Any]=4096 , UpperCAmelCase__ : int=16 , UpperCAmelCase__ : Tuple=12 , UpperCAmelCase__ : int=4096 , UpperCAmelCase__ : List[Any]=16 , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : Tuple="gelu" , UpperCAmelCase__ : Union[str, Any]=1024 , UpperCAmelCase__ : List[str]=0.1 , UpperCAmelCase__ : Any=0.0 , UpperCAmelCase__ : Dict=0.0 , UpperCAmelCase__ : Tuple=0.0_2 , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : Optional[Any]=False , UpperCAmelCase__ : int=True , UpperCAmelCase__ : Tuple=1 , UpperCAmelCase__ : Dict=0 , UpperCAmelCase__ : Union[str, Any]=2 , UpperCAmelCase__ : Optional[int]=True , UpperCAmelCase__ : Tuple=2 , UpperCAmelCase__ : Any=2 , UpperCAmelCase__ : Optional[Any]=False , UpperCAmelCase__ : Dict=100 , UpperCAmelCase__ : Union[str, Any]=800 , **UpperCAmelCase__ : Dict , ) -> List[Any]: _a : Any = vocab_size _a : Any = max_position_embeddings _a : Union[str, Any] = d_model _a : List[str] = encoder_ffn_dim _a : List[Any] = encoder_layers _a : Dict = encoder_attention_heads _a : Tuple = decoder_ffn_dim _a : List[Any] = decoder_layers _a : Optional[Any] = decoder_attention_heads _a : Optional[Any] = dropout _a : str = attention_dropout _a : Dict = activation_dropout _a : Any = activation_function _a : Tuple = init_std _a : Dict = encoder_layerdrop _a : Optional[int] = decoder_layerdrop _a : Optional[Any] = classifier_dropout _a : List[Any] = use_cache _a : Dict = encoder_layers _a : str = scale_embedding # scale factor will be sqrt(d_model) if True _a : int = use_prompt _a : Dict = prompt_length _a : Dict = prompt_mid_dim super().__init__( pad_token_id=UpperCAmelCase__ , bos_token_id=UpperCAmelCase__ , eos_token_id=UpperCAmelCase__ , is_encoder_decoder=UpperCAmelCase__ , decoder_start_token_id=UpperCAmelCase__ , forced_eos_token_id=UpperCAmelCase__ , **UpperCAmelCase__ , ) if self.forced_bos_token_id is None and kwargs.get("""force_bos_token_to_be_generated""" , UpperCAmelCase__ ): _a : List[str] = self.bos_token_id warnings.warn( f"""Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. """ """The config can simply be saved and uploaded again to be fixed.""" )
294
1
"""simple docstring""" from __future__ import annotations from PIL import Image # Define glider example _snake_case = [ [0, 1, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], ] # Define blinker example _snake_case = [[0, 1, 0], [0, 1, 0], [0, 1, 0]] def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : List[Any] = [] for i in range(len(UpperCamelCase__ ) ): _a : Any = [] for j in range(len(cells[i] ) ): # Get the number of live neighbours _a : Optional[Any] = 0 if i > 0 and j > 0: neighbour_count += cells[i - 1][j - 1] if i > 0: neighbour_count += cells[i - 1][j] if i > 0 and j < len(cells[i] ) - 1: neighbour_count += cells[i - 1][j + 1] if j > 0: neighbour_count += cells[i][j - 1] if j < len(cells[i] ) - 1: neighbour_count += cells[i][j + 1] if i < len(UpperCamelCase__ ) - 1 and j > 0: neighbour_count += cells[i + 1][j - 1] if i < len(UpperCamelCase__ ) - 1: neighbour_count += cells[i + 1][j] if i < len(UpperCamelCase__ ) - 1 and j < len(cells[i] ) - 1: neighbour_count += cells[i + 1][j + 1] # Rules of the game of life (excerpt from Wikipedia): # 1. Any live cell with two or three live neighbours survives. # 2. Any dead cell with three live neighbours becomes a live cell. # 3. All other live cells die in the next generation. # Similarly, all other dead cells stay dead. _a : Optional[int] = cells[i][j] == 1 if ( (alive and 2 <= neighbour_count <= 3) or not alive and neighbour_count == 3 ): next_generation_row.append(1 ) else: next_generation_row.append(0 ) next_generation.append(UpperCamelCase__ ) return next_generation def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : List[Any] = [] for _ in range(UpperCamelCase__ ): # Create output image _a : int = Image.new("""RGB""" , (len(cells[0] ), len(UpperCamelCase__ )) ) _a : Optional[Any] = img.load() # Save cells to image for x in range(len(UpperCamelCase__ ) ): for y in range(len(cells[0] ) ): _a : List[Any] = 2_5_5 - cells[y][x] * 2_5_5 _a : Tuple = (colour, colour, colour) # Save image images.append(UpperCamelCase__ ) _a : Tuple = new_generation(UpperCamelCase__ ) return images if __name__ == "__main__": _snake_case = generate_images(GLIDER, 16) images[0].save('out.gif', save_all=True, append_images=images[1:])
294
"""simple docstring""" import logging import os from typing import List, TextIO, Union from conllu import parse_incr from utils_ner import InputExample, Split, TokenClassificationTask _snake_case = logging.getLogger(__name__) class UpperCamelCase ( snake_case_ ): def __init__( self : Optional[Any] , UpperCAmelCase__ : Optional[int]=-1 ) -> Tuple: # in NER datasets, the last column is usually reserved for NER label _a : Optional[int] = label_idx def _lowercase ( self : Any , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Union[Split, str] ) -> List[InputExample]: if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): _a : Any = mode.value _a : Optional[int] = os.path.join(UpperCAmelCase__ , f"""{mode}.txt""" ) _a : int = 1 _a : int = [] with open(UpperCAmelCase__ , encoding="""utf-8""" ) as f: _a : str = [] _a : str = [] for line in f: if line.startswith("""-DOCSTART-""" ) or line == "" or line == "\n": if words: examples.append(InputExample(guid=f"""{mode}-{guid_index}""" , words=UpperCAmelCase__ , labels=UpperCAmelCase__ ) ) guid_index += 1 _a : List[str] = [] _a : str = [] else: _a : List[Any] = line.split(""" """ ) words.append(splits[0] ) if len(UpperCAmelCase__ ) > 1: labels.append(splits[self.label_idx].replace("""\n""" , """""" ) ) else: # Examples could have no label for mode = "test" labels.append("""O""" ) if words: examples.append(InputExample(guid=f"""{mode}-{guid_index}""" , words=UpperCAmelCase__ , labels=UpperCAmelCase__ ) ) return examples def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : List ) -> Union[str, Any]: _a : List[str] = 0 for line in test_input_reader: if line.startswith("""-DOCSTART-""" ) or line == "" or line == "\n": writer.write(UpperCAmelCase__ ) if not preds_list[example_id]: example_id += 1 elif preds_list[example_id]: _a : int = line.split()[0] + """ """ + preds_list[example_id].pop(0 ) + """\n""" writer.write(UpperCAmelCase__ ) else: logger.warning("""Maximum sequence length exceeded: No prediction for '%s'.""" , line.split()[0] ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : str ) -> List[str]: if path: with open(UpperCAmelCase__ , """r""" ) as f: _a : List[Any] = f.read().splitlines() if "O" not in labels: _a : Union[str, Any] = ["""O"""] + labels return labels else: return ["O", "B-MISC", "I-MISC", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC"] class UpperCamelCase ( snake_case_ ): def __init__( self : Union[str, Any] ) -> List[str]: # in CONLL2003 dataset chunk column is second-to-last super().__init__(label_idx=-2 ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : str ) -> List[str]: if path: with open(UpperCAmelCase__ , """r""" ) as f: _a : Optional[int] = f.read().splitlines() if "O" not in labels: _a : Optional[Any] = ["""O"""] + labels return labels else: return [ "O", "B-ADVP", "B-INTJ", "B-LST", "B-PRT", "B-NP", "B-SBAR", "B-VP", "B-ADJP", "B-CONJP", "B-PP", "I-ADVP", "I-INTJ", "I-LST", "I-PRT", "I-NP", "I-SBAR", "I-VP", "I-ADJP", "I-CONJP", "I-PP", ] class UpperCamelCase ( snake_case_ ): def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Union[Split, str] ) -> List[InputExample]: if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): _a : List[Any] = mode.value _a : Union[str, Any] = os.path.join(UpperCAmelCase__ , f"""{mode}.txt""" ) _a : List[str] = 1 _a : Optional[Any] = [] with open(UpperCAmelCase__ , encoding="""utf-8""" ) as f: for sentence in parse_incr(UpperCAmelCase__ ): _a : List[Any] = [] _a : Any = [] for token in sentence: words.append(token["""form"""] ) labels.append(token["""upos"""] ) assert len(UpperCAmelCase__ ) == len(UpperCAmelCase__ ) if words: examples.append(InputExample(guid=f"""{mode}-{guid_index}""" , words=UpperCAmelCase__ , labels=UpperCAmelCase__ ) ) guid_index += 1 return examples def _lowercase ( self : Tuple , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : List ) -> Dict: _a : Optional[Any] = 0 for sentence in parse_incr(UpperCAmelCase__ ): _a : List[str] = preds_list[example_id] _a : str = """""" for token in sentence: out += f"""{token['form']} ({token['upos']}|{s_p.pop(0 )}) """ out += "\n" writer.write(UpperCAmelCase__ ) example_id += 1 def _lowercase ( self : List[str] , UpperCAmelCase__ : str ) -> List[str]: if path: with open(UpperCAmelCase__ , """r""" ) as f: return f.read().splitlines() else: return [ "ADJ", "ADP", "ADV", "AUX", "CCONJ", "DET", "INTJ", "NOUN", "NUM", "PART", "PRON", "PROPN", "PUNCT", "SCONJ", "SYM", "VERB", "X", ]
294
1
"""simple docstring""" import math import sys def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' if number != int(UpperCamelCase__ ): raise ValueError("""the value of input must be a natural number""" ) if number < 0: raise ValueError("""the value of input must not be a negative number""" ) if number == 0: return 1 _a : str = [-1] * (number + 1) _a : List[Any] = 0 for i in range(1 , number + 1 ): _a : List[str] = sys.maxsize _a : str = int(math.sqrt(UpperCamelCase__ ) ) for j in range(1 , root + 1 ): _a : Union[str, Any] = 1 + answers[i - (j**2)] _a : Optional[int] = min(UpperCamelCase__ , UpperCamelCase__ ) _a : List[str] = answer return answers[number] if __name__ == "__main__": import doctest doctest.testmod()
294
"""simple docstring""" from __future__ import annotations import time import numpy as np _snake_case = [8, 5, 9, 7] _snake_case = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] _snake_case = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class UpperCamelCase : def __init__( self : List[Any] , UpperCAmelCase__ : list[int] , UpperCAmelCase__ : list[list[int]] , UpperCAmelCase__ : list[list[int]] , ) -> None: _a : List[str] = claim_vector _a : List[Any] = allocated_resources_table _a : Union[str, Any] = maximum_claim_table def _lowercase ( self : Tuple ) -> list[int]: return [ sum(p_item[i] for p_item in self.__allocated_resources_table ) for i in range(len(self.__allocated_resources_table[0] ) ) ] def _lowercase ( self : int ) -> list[int]: return np.array(self.__claim_vector ) - np.array( self.__processes_resource_summation() ) def _lowercase ( self : List[str] ) -> list[list[int]]: return [ list(np.array(self.__maximum_claim_table[i] ) - np.array(UpperCAmelCase__ ) ) for i, allocated_resource in enumerate(self.__allocated_resources_table ) ] def _lowercase ( self : Optional[Any] ) -> dict[int, list[int]]: return {self.__need().index(UpperCAmelCase__ ): i for i in self.__need()} def _lowercase ( self : Dict , **UpperCAmelCase__ : Optional[Any] ) -> None: _a : List[Any] = self.__need() _a : Optional[int] = self.__allocated_resources_table _a : str = self.__available_resources() _a : Optional[Any] = self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print("""_""" * 50 + """\n""" ) while need_list: _a : int = False for each_need in need_list: _a : Optional[int] = True for index, need in enumerate(UpperCAmelCase__ ): if need > available_resources[index]: _a : List[Any] = False break if execution: _a : str = True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: _a : Any = original_need_index print(f"""Process {process_number + 1} is executing.""" ) # remove the process run from stack need_list.remove(UpperCAmelCase__ ) # update available/freed resources stack _a : Union[str, Any] = np.array(UpperCAmelCase__ ) + np.array( alloc_resources_table[process_number] ) print( """Updated available resource stack for processes: """ + """ """.join([str(UpperCAmelCase__ ) for x in available_resources] ) ) break if safe: print("""The process is in a safe state.\n""" ) else: print("""System in unsafe state. Aborting...\n""" ) break def _lowercase ( self : Any ) -> Optional[int]: print(""" """ * 9 + """Allocated Resource Table""" ) for item in self.__allocated_resources_table: print( f"""P{self.__allocated_resources_table.index(UpperCAmelCase__ ) + 1}""" + """ """.join(f"""{it:>8}""" for it in item ) + """\n""" ) print(""" """ * 9 + """System Resource Table""" ) for item in self.__maximum_claim_table: print( f"""P{self.__maximum_claim_table.index(UpperCAmelCase__ ) + 1}""" + """ """.join(f"""{it:>8}""" for it in item ) + """\n""" ) print( """Current Usage by Active Processes: """ + """ """.join(str(UpperCAmelCase__ ) for x in self.__claim_vector ) ) print( """Initial Available Resources: """ + """ """.join(str(UpperCAmelCase__ ) for x in self.__available_resources() ) ) time.sleep(1 ) if __name__ == "__main__": import doctest doctest.testmod()
294
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { 'vinvino02/glpn-kitti': 'https://huggingface.co/vinvino02/glpn-kitti/resolve/main/config.json', # See all GLPN models at https://huggingface.co/models?filter=glpn } class UpperCamelCase ( snake_case_ ): UpperCamelCase : str = '''glpn''' def __init__( self : List[Any] , UpperCAmelCase__ : List[str]=3 , UpperCAmelCase__ : Dict=4 , UpperCAmelCase__ : str=[2, 2, 2, 2] , UpperCAmelCase__ : Tuple=[8, 4, 2, 1] , UpperCAmelCase__ : Optional[Any]=[32, 64, 160, 256] , UpperCAmelCase__ : Dict=[7, 3, 3, 3] , UpperCAmelCase__ : int=[4, 2, 2, 2] , UpperCAmelCase__ : Optional[int]=[1, 2, 5, 8] , UpperCAmelCase__ : List[str]=[4, 4, 4, 4] , UpperCAmelCase__ : Union[str, Any]="gelu" , UpperCAmelCase__ : List[str]=0.0 , UpperCAmelCase__ : Any=0.0 , UpperCAmelCase__ : List[str]=0.0_2 , UpperCAmelCase__ : Union[str, Any]=0.1 , UpperCAmelCase__ : Optional[Any]=1E-6 , UpperCAmelCase__ : Tuple=64 , UpperCAmelCase__ : Optional[Any]=10 , UpperCAmelCase__ : Union[str, Any]=-1 , **UpperCAmelCase__ : Union[str, Any] , ) -> str: super().__init__(**UpperCAmelCase__ ) _a : int = num_channels _a : Dict = num_encoder_blocks _a : int = depths _a : List[Any] = sr_ratios _a : Dict = hidden_sizes _a : Any = patch_sizes _a : Union[str, Any] = strides _a : List[Any] = mlp_ratios _a : Optional[Any] = num_attention_heads _a : str = hidden_act _a : Tuple = hidden_dropout_prob _a : Optional[Any] = attention_probs_dropout_prob _a : int = initializer_range _a : Union[str, Any] = drop_path_rate _a : Optional[int] = layer_norm_eps _a : Any = decoder_hidden_size _a : Union[str, Any] = max_depth _a : List[Any] = head_in_index
294
"""simple docstring""" from collections.abc import Iterable from typing import Generic, TypeVar _snake_case = TypeVar('_T') class UpperCamelCase ( Generic[_T] ): def __init__( self : Optional[int] , UpperCAmelCase__ : Iterable[_T] | None = None ) -> None: _a : list[_T] = list(iterable or [] ) _a : list[_T] = [] def __len__( self : str ) -> int: return len(self._stacka ) + len(self._stacka ) def __repr__( self : List[str] ) -> str: return f"""Queue({tuple(self._stacka[::-1] + self._stacka )})""" def _lowercase ( self : Union[str, Any] , UpperCAmelCase__ : _T ) -> None: self._stacka.append(UpperCAmelCase__ ) def _lowercase ( self : Optional[Any] ) -> _T: _a : Any = self._stacka.pop _a : Union[str, Any] = self._stacka.append if not self._stacka: while self._stacka: stacka_append(stacka_pop() ) if not self._stacka: raise IndexError("""Queue is empty""" ) return self._stacka.pop() if __name__ == "__main__": from doctest import testmod testmod()
294
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case = {'configuration_ibert': ['IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'IBertConfig', 'IBertOnnxConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'IBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'IBertForMaskedLM', 'IBertForMultipleChoice', 'IBertForQuestionAnswering', 'IBertForSequenceClassification', 'IBertForTokenClassification', 'IBertModel', 'IBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
294
"""simple docstring""" import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device _snake_case = False class UpperCamelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : Optional[Any] ) -> Optional[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self : Tuple ) -> List[Any]: _a : Tuple = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : List[str] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) _a : Optional[Any] = torch.manual_seed(0 ) _a : Union[str, Any] = pipe.dual_guided( prompt="""first prompt""" , image=UpperCAmelCase__ , text_to_image_strength=0.7_5 , generator=UpperCAmelCase__ , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(UpperCAmelCase__ ) _a : Dict = VersatileDiffusionPipeline.from_pretrained(UpperCAmelCase__ , torch_dtype=torch.floataa ) pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Optional[Any] = generator.manual_seed(0 ) _a : str = pipe.dual_guided( prompt="""first prompt""" , image=UpperCAmelCase__ , text_to_image_strength=0.7_5 , generator=UpperCAmelCase__ , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def _lowercase ( self : Optional[int] ) -> Optional[int]: _a : Optional[int] = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : int = """cyberpunk 2077""" _a : Tuple = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) _a : Tuple = torch.manual_seed(0 ) _a : Any = pipe.dual_guided( prompt=UpperCAmelCase__ , image=UpperCAmelCase__ , text_to_image_strength=0.7_5 , generator=UpperCAmelCase__ , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" , ).images _a : List[str] = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _a : Optional[int] = np.array([0.1_4_4_8, 0.1_6_1_9, 0.1_7_4_1, 0.1_0_8_6, 0.1_1_4_7, 0.1_1_2_8, 0.1_1_9_9, 0.1_1_6_5, 0.1_0_0_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 _a : int = """A painting of a squirrel eating a burger """ _a : Tuple = torch.manual_seed(0 ) _a : Union[str, Any] = pipe.text_to_image( prompt=UpperCAmelCase__ , generator=UpperCAmelCase__ , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" ).images _a : int = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _a : int = np.array([0.3_3_6_7, 0.3_1_6_9, 0.2_6_5_6, 0.3_8_7_0, 0.4_7_9_0, 0.3_7_9_6, 0.4_0_0_9, 0.4_8_7_8, 0.4_7_7_8] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 _a : str = pipe.image_variation(UpperCAmelCase__ , generator=UpperCAmelCase__ , output_type="""numpy""" ).images _a : str = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _a : Optional[Any] = np.array([0.3_0_7_6, 0.3_1_2_3, 0.3_2_8_4, 0.3_7_8_2, 0.3_7_7_0, 0.3_8_9_4, 0.4_2_9_7, 0.4_3_3_1, 0.4_4_5_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1
294
1
"""simple docstring""" def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
294
"""simple docstring""" import json import os from typing import Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = {'vocab_file': 'vocab.json'} _snake_case = { 'vocab_file': { 'mgp-str': 'https://huggingface.co/alibaba-damo/mgp-str-base/blob/main/vocab.json', } } _snake_case = {'mgp-str': 27} class UpperCamelCase ( snake_case_ ): UpperCamelCase : List[str] = VOCAB_FILES_NAMES UpperCamelCase : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : Optional[int] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : List[Any]="[GO]" , UpperCAmelCase__ : Tuple="[GO]" , UpperCAmelCase__ : Optional[int]="[s]" , UpperCAmelCase__ : int="[GO]" , **UpperCAmelCase__ : Dict ) -> int: super().__init__( unk_token=UpperCAmelCase__ , bos_token=UpperCAmelCase__ , eos_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , **UpperCAmelCase__ , ) with open(UpperCAmelCase__ , encoding="""utf-8""" ) as vocab_handle: _a : int = json.load(UpperCAmelCase__ ) _a : Optional[int] = {v: k for k, v in self.vocab.items()} @property def _lowercase ( self : Dict ) -> Union[str, Any]: return len(self.vocab ) def _lowercase ( self : Union[str, Any] ) -> str: return dict(self.vocab , **self.added_tokens_encoder ) def _lowercase ( self : Dict , UpperCAmelCase__ : str ) -> Union[str, Any]: _a : Tuple = [] for s in text: char_tokens.extend(UpperCAmelCase__ ) return char_tokens def _lowercase ( self : List[Any] , UpperCAmelCase__ : str ) -> Dict: return self.vocab.get(UpperCAmelCase__ , self.vocab.get(self.unk_token ) ) def _lowercase ( self : Optional[int] , UpperCAmelCase__ : Tuple ) -> List[Any]: return self.decoder.get(UpperCAmelCase__ ) def _lowercase ( self : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(UpperCAmelCase__ ): logger.error("""Vocabulary path ({}) should be a directory""".format(UpperCAmelCase__ ) ) return _a : Tuple = os.path.join( UpperCAmelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) with open(UpperCAmelCase__ , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.vocab , indent=2 , sort_keys=UpperCAmelCase__ , ensure_ascii=UpperCAmelCase__ ) + """\n""" ) return (vocab_file,)
294
1
"""simple docstring""" import torch from diffusers import DDPMParallelScheduler from .test_schedulers import SchedulerCommonTest class UpperCamelCase ( snake_case_ ): UpperCamelCase : Dict = (DDPMParallelScheduler,) def _lowercase ( self : Optional[int] , **UpperCAmelCase__ : str ) -> List[str]: _a : Optional[int] = { """num_train_timesteps""": 1000, """beta_start""": 0.0_0_0_1, """beta_end""": 0.0_2, """beta_schedule""": """linear""", """variance_type""": """fixed_small""", """clip_sample""": True, } config.update(**UpperCAmelCase__ ) return config def _lowercase ( self : Any ) -> Union[str, Any]: for timesteps in [1, 5, 100, 1000]: self.check_over_configs(num_train_timesteps=UpperCAmelCase__ ) def _lowercase ( self : List[Any] ) -> List[str]: for beta_start, beta_end in zip([0.0_0_0_1, 0.0_0_1, 0.0_1, 0.1] , [0.0_0_2, 0.0_2, 0.2, 2] ): self.check_over_configs(beta_start=UpperCAmelCase__ , beta_end=UpperCAmelCase__ ) def _lowercase ( self : Any ) -> Dict: for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=UpperCAmelCase__ ) def _lowercase ( self : str ) -> int: for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=UpperCAmelCase__ ) def _lowercase ( self : Union[str, Any] ) -> Union[str, Any]: for clip_sample in [True, False]: self.check_over_configs(clip_sample=UpperCAmelCase__ ) def _lowercase ( self : Tuple ) -> Any: self.check_over_configs(thresholding=UpperCAmelCase__ ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=UpperCAmelCase__ , prediction_type=UpperCAmelCase__ , sample_max_value=UpperCAmelCase__ , ) def _lowercase ( self : Optional[Any] ) -> Union[str, Any]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=UpperCAmelCase__ ) def _lowercase ( self : Optional[int] ) -> Optional[Any]: for t in [0, 500, 999]: self.check_over_forward(time_step=UpperCAmelCase__ ) def _lowercase ( self : Union[str, Any] ) -> Tuple: _a : int = self.scheduler_classes[0] _a : str = self.get_scheduler_config() _a : int = scheduler_class(**UpperCAmelCase__ ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0_0_9_7_9 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.0_2 ) ) < 1E-5 def _lowercase ( self : str ) -> List[str]: _a : List[str] = self.scheduler_classes[0] _a : List[str] = self.get_scheduler_config() _a : Union[str, Any] = scheduler_class(**UpperCAmelCase__ ) _a : Dict = len(UpperCAmelCase__ ) _a : int = self.dummy_model() _a : List[Any] = self.dummy_sample_deter _a : str = self.dummy_sample_deter + 0.1 _a : List[Any] = self.dummy_sample_deter - 0.1 _a : List[Any] = samplea.shape[0] _a : Tuple = torch.stack([samplea, samplea, samplea] , dim=0 ) _a : Dict = torch.arange(UpperCAmelCase__ )[0:3, None].repeat(1 , UpperCAmelCase__ ) _a : Optional[int] = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) _a : Optional[Any] = scheduler.batch_step_no_noise(UpperCAmelCase__ , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) ) _a : List[str] = torch.sum(torch.abs(UpperCAmelCase__ ) ) _a : int = torch.mean(torch.abs(UpperCAmelCase__ ) ) assert abs(result_sum.item() - 1_1_5_3.1_8_3_3 ) < 1E-2 assert abs(result_mean.item() - 0.5_0_0_5 ) < 1E-3 def _lowercase ( self : List[Any] ) -> Union[str, Any]: _a : Dict = self.scheduler_classes[0] _a : Optional[Any] = self.get_scheduler_config() _a : Optional[int] = scheduler_class(**UpperCAmelCase__ ) _a : Optional[Any] = len(UpperCAmelCase__ ) _a : Union[str, Any] = self.dummy_model() _a : int = self.dummy_sample_deter _a : List[str] = torch.manual_seed(0 ) for t in reversed(range(UpperCAmelCase__ ) ): # 1. predict noise residual _a : str = model(UpperCAmelCase__ , UpperCAmelCase__ ) # 2. predict previous mean of sample x_t-1 _a : Union[str, Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , generator=UpperCAmelCase__ ).prev_sample _a : List[str] = pred_prev_sample _a : List[Any] = torch.sum(torch.abs(UpperCAmelCase__ ) ) _a : Dict = torch.mean(torch.abs(UpperCAmelCase__ ) ) assert abs(result_sum.item() - 2_5_8.9_6_0_6 ) < 1E-2 assert abs(result_mean.item() - 0.3_3_7_2 ) < 1E-3 def _lowercase ( self : str ) -> int: _a : List[Any] = self.scheduler_classes[0] _a : List[Any] = self.get_scheduler_config(prediction_type="""v_prediction""" ) _a : int = scheduler_class(**UpperCAmelCase__ ) _a : Dict = len(UpperCAmelCase__ ) _a : List[str] = self.dummy_model() _a : List[str] = self.dummy_sample_deter _a : str = torch.manual_seed(0 ) for t in reversed(range(UpperCAmelCase__ ) ): # 1. predict noise residual _a : Dict = model(UpperCAmelCase__ , UpperCAmelCase__ ) # 2. predict previous mean of sample x_t-1 _a : List[str] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , generator=UpperCAmelCase__ ).prev_sample _a : Any = pred_prev_sample _a : List[str] = torch.sum(torch.abs(UpperCAmelCase__ ) ) _a : Tuple = torch.mean(torch.abs(UpperCAmelCase__ ) ) assert abs(result_sum.item() - 2_0_2.0_2_9_6 ) < 1E-2 assert abs(result_mean.item() - 0.2_6_3_1 ) < 1E-3 def _lowercase ( self : Union[str, Any] ) -> Dict: _a : int = self.scheduler_classes[0] _a : List[Any] = self.get_scheduler_config() _a : Optional[Any] = scheduler_class(**UpperCAmelCase__ ) _a : Dict = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=UpperCAmelCase__ ) _a : Any = scheduler.timesteps for i, timestep in enumerate(UpperCAmelCase__ ): if i == len(UpperCAmelCase__ ) - 1: _a : Dict = -1 else: _a : Tuple = timesteps[i + 1] _a : Union[str, Any] = scheduler.previous_timestep(UpperCAmelCase__ ) _a : Tuple = prev_t.item() self.assertEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def _lowercase ( self : Any ) -> Optional[int]: _a : Dict = self.scheduler_classes[0] _a : Optional[Any] = self.get_scheduler_config() _a : Optional[Any] = scheduler_class(**UpperCAmelCase__ ) _a : Union[str, Any] = [100, 87, 50, 51, 0] with self.assertRaises(UpperCAmelCase__ , msg="""`custom_timesteps` must be in descending order.""" ): scheduler.set_timesteps(timesteps=UpperCAmelCase__ ) def _lowercase ( self : Dict ) -> List[Any]: _a : Dict = self.scheduler_classes[0] _a : int = self.get_scheduler_config() _a : List[str] = scheduler_class(**UpperCAmelCase__ ) _a : Dict = [100, 87, 50, 1, 0] _a : Optional[int] = len(UpperCAmelCase__ ) with self.assertRaises(UpperCAmelCase__ , msg="""Can only pass one of `num_inference_steps` or `custom_timesteps`.""" ): scheduler.set_timesteps(num_inference_steps=UpperCAmelCase__ , timesteps=UpperCAmelCase__ ) def _lowercase ( self : Union[str, Any] ) -> Any: _a : Union[str, Any] = self.scheduler_classes[0] _a : Optional[Any] = self.get_scheduler_config() _a : Union[str, Any] = scheduler_class(**UpperCAmelCase__ ) _a : int = [scheduler.config.num_train_timesteps] with self.assertRaises( UpperCAmelCase__ , msg="""`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}""" , ): scheduler.set_timesteps(timesteps=UpperCAmelCase__ )
294
"""simple docstring""" import tempfile import torch from diffusers import IPNDMScheduler from .test_schedulers import SchedulerCommonTest class UpperCamelCase ( snake_case_ ): UpperCamelCase : int = (IPNDMScheduler,) UpperCamelCase : int = (('''num_inference_steps''', 50),) def _lowercase ( self : Union[str, Any] , **UpperCAmelCase__ : Tuple ) -> int: _a : Optional[int] = {"""num_train_timesteps""": 1000} config.update(**UpperCAmelCase__ ) return config def _lowercase ( self : Dict , UpperCAmelCase__ : Any=0 , **UpperCAmelCase__ : Optional[Any] ) -> Union[str, Any]: _a : Optional[int] = dict(self.forward_default_kwargs ) _a : Dict = kwargs.pop("""num_inference_steps""" , UpperCAmelCase__ ) _a : Optional[Any] = self.dummy_sample _a : Union[str, Any] = 0.1 * sample _a : Union[str, Any] = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] for scheduler_class in self.scheduler_classes: _a : Optional[int] = self.get_scheduler_config(**UpperCAmelCase__ ) _a : Union[str, Any] = scheduler_class(**UpperCAmelCase__ ) scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residuals _a : Any = dummy_past_residuals[:] if time_step is None: _a : str = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(UpperCAmelCase__ ) _a : Union[str, Any] = scheduler_class.from_pretrained(UpperCAmelCase__ ) new_scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residuals _a : Optional[Any] = dummy_past_residuals[:] _a : List[Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : str = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" _a : Optional[int] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : Tuple = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def _lowercase ( self : Tuple ) -> List[str]: pass def _lowercase ( self : Optional[int] , UpperCAmelCase__ : List[str]=0 , **UpperCAmelCase__ : Optional[Any] ) -> List[Any]: _a : Optional[Any] = dict(self.forward_default_kwargs ) _a : Optional[Any] = kwargs.pop("""num_inference_steps""" , UpperCAmelCase__ ) _a : Optional[Any] = self.dummy_sample _a : List[Any] = 0.1 * sample _a : Optional[Any] = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] for scheduler_class in self.scheduler_classes: _a : Union[str, Any] = self.get_scheduler_config() _a : Optional[Any] = scheduler_class(**UpperCAmelCase__ ) scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residuals (must be after setting timesteps) _a : Any = dummy_past_residuals[:] if time_step is None: _a : List[Any] = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(UpperCAmelCase__ ) _a : Any = scheduler_class.from_pretrained(UpperCAmelCase__ ) # copy over dummy past residuals new_scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residual (must be after setting timesteps) _a : Optional[Any] = dummy_past_residuals[:] _a : List[str] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : Tuple = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" _a : Union[str, Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : int = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def _lowercase ( self : str , **UpperCAmelCase__ : Any ) -> List[str]: _a : Optional[int] = self.scheduler_classes[0] _a : Optional[Any] = self.get_scheduler_config(**UpperCAmelCase__ ) _a : Union[str, Any] = scheduler_class(**UpperCAmelCase__ ) _a : int = 10 _a : List[Any] = self.dummy_model() _a : str = self.dummy_sample_deter scheduler.set_timesteps(UpperCAmelCase__ ) for i, t in enumerate(scheduler.timesteps ): _a : str = model(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ).prev_sample for i, t in enumerate(scheduler.timesteps ): _a : Union[str, Any] = model(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Any = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ).prev_sample return sample def _lowercase ( self : int ) -> str: _a : Dict = dict(self.forward_default_kwargs ) _a : int = kwargs.pop("""num_inference_steps""" , UpperCAmelCase__ ) for scheduler_class in self.scheduler_classes: _a : Optional[int] = self.get_scheduler_config() _a : Tuple = scheduler_class(**UpperCAmelCase__ ) _a : Tuple = self.dummy_sample _a : Optional[Any] = 0.1 * sample if num_inference_steps is not None and hasattr(UpperCAmelCase__ , """set_timesteps""" ): scheduler.set_timesteps(UpperCAmelCase__ ) elif num_inference_steps is not None and not hasattr(UpperCAmelCase__ , """set_timesteps""" ): _a : List[str] = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) _a : Union[str, Any] = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] _a : Optional[Any] = dummy_past_residuals[:] _a : Optional[Any] = scheduler.timesteps[5] _a : str = scheduler.timesteps[6] _a : Optional[int] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : Union[str, Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) _a : Tuple = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : List[str] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def _lowercase ( self : List[str] ) -> List[str]: for timesteps in [100, 1000]: self.check_over_configs(num_train_timesteps=UpperCAmelCase__ , time_step=UpperCAmelCase__ ) def _lowercase ( self : List[str] ) -> List[str]: for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 100] ): self.check_over_forward(num_inference_steps=UpperCAmelCase__ , time_step=UpperCAmelCase__ ) def _lowercase ( self : int ) -> List[Any]: _a : str = self.full_loop() _a : List[Any] = torch.mean(torch.abs(UpperCAmelCase__ ) ) assert abs(result_mean.item() - 2540529 ) < 10
294
1
"""simple docstring""" import os import unittest from transformers.models.bartpho.tokenization_bartpho import VOCAB_FILES_NAMES, BartphoTokenizer from transformers.testing_utils import get_tests_dir from ...test_tokenization_common import TokenizerTesterMixin _snake_case = get_tests_dir('fixtures/test_sentencepiece_bpe.model') class UpperCamelCase ( snake_case_ , unittest.TestCase ): UpperCamelCase : Tuple = BartphoTokenizer UpperCamelCase : Dict = False UpperCamelCase : Union[str, Any] = True def _lowercase ( self : Any ) -> List[Any]: super().setUp() _a : List[Any] = ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] _a : Any = dict(zip(UpperCAmelCase__ , range(len(UpperCAmelCase__ ) ) ) ) _a : Optional[int] = {"""unk_token""": """<unk>"""} _a : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""monolingual_vocab_file"""] ) with open(self.monolingual_vocab_file , """w""" , encoding="""utf-8""" ) as fp: for token in vocab_tokens: fp.write(f"""{token} {vocab_tokens[token]}\n""" ) _a : List[str] = BartphoTokenizer(UpperCAmelCase__ , self.monolingual_vocab_file , **self.special_tokens_map ) tokenizer.save_pretrained(self.tmpdirname ) def _lowercase ( self : List[str] , **UpperCAmelCase__ : str ) -> Union[str, Any]: kwargs.update(self.special_tokens_map ) return BartphoTokenizer.from_pretrained(self.tmpdirname , **UpperCAmelCase__ ) def _lowercase ( self : str , UpperCAmelCase__ : Optional[Any] ) -> Union[str, Any]: _a : Dict = """This is a là test""" _a : str = """This is a<unk><unk> test""" return input_text, output_text def _lowercase ( self : Union[str, Any] ) -> Optional[int]: _a : List[str] = BartphoTokenizer(UpperCAmelCase__ , self.monolingual_vocab_file , **self.special_tokens_map ) _a : List[str] = """This is a là test""" _a : Optional[Any] = """▁This ▁is ▁a ▁l à ▁t est""".split() _a : Dict = tokenizer.tokenize(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Tuple = tokens + [tokenizer.unk_token] _a : List[Any] = [4, 5, 6, 3, 3, 7, 8, 3] self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) , UpperCAmelCase__ )
294
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging _snake_case = logging.get_logger(__name__) _snake_case = {'vocab_file': 'spiece.model'} _snake_case = { 'vocab_file': { 'TsinghuaAI/CPM-Generate': 'https://huggingface.co/TsinghuaAI/CPM-Generate/resolve/main/spiece.model', } } class UpperCamelCase ( snake_case_ ): def __init__( self : Tuple , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : int=False , UpperCAmelCase__ : Dict=True , UpperCAmelCase__ : Union[str, Any]=False , UpperCAmelCase__ : Dict="<s>" , UpperCAmelCase__ : Any="</s>" , UpperCAmelCase__ : Any="<unk>" , UpperCAmelCase__ : int="<sep>" , UpperCAmelCase__ : Tuple="<pad>" , UpperCAmelCase__ : Any="<cls>" , UpperCAmelCase__ : Optional[Any]="<mask>" , UpperCAmelCase__ : int=["<eop>", "<eod>"] , UpperCAmelCase__ : Optional[Dict[str, Any]] = None , **UpperCAmelCase__ : List[str] , ) -> None: _a : Optional[int] = AddedToken(UpperCAmelCase__ , lstrip=UpperCAmelCase__ , rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else mask_token _a : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=UpperCAmelCase__ , remove_space=UpperCAmelCase__ , keep_accents=UpperCAmelCase__ , bos_token=UpperCAmelCase__ , eos_token=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , sep_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , cls_token=UpperCAmelCase__ , mask_token=UpperCAmelCase__ , additional_special_tokens=UpperCAmelCase__ , sp_model_kwargs=self.sp_model_kwargs , **UpperCAmelCase__ , ) _a : Optional[Any] = 3 _a : Tuple = do_lower_case _a : Tuple = remove_space _a : Tuple = keep_accents _a : Tuple = vocab_file _a : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCAmelCase__ ) try: import jieba except ModuleNotFoundError as error: raise error.__class__( """You need to install jieba to use CpmTokenizer or CpmTokenizerFast. """ """See https://pypi.org/project/jieba/ for installation.""" ) _a : int = jieba _a : Tuple = str.maketrans(""" \n""" , """\u2582\u2583""" ) @property # Copied from transformers.models.xlnet.tokenization_xlnet.XLNetTokenizer.vocab_size def _lowercase ( self : Optional[Any] ) -> Any: return len(self.sp_model ) def _lowercase ( self : str ) -> Union[str, Any]: _a : int = {self.convert_ids_to_tokens(UpperCAmelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Tuple ) -> List[str]: _a : Tuple = self.__dict__.copy() _a : Tuple = None return state def __setstate__( self : Any , UpperCAmelCase__ : Dict ) -> Dict: _a : Tuple = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): _a : Tuple = {} _a : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _lowercase ( self : List[str] , UpperCAmelCase__ : Union[str, Any] ) -> Dict: if self.remove_space: _a : Optional[int] = """ """.join(inputs.strip().split() ) else: _a : List[Any] = inputs _a : int = outputs.replace("""``""" , """\"""" ).replace("""''""" , """\"""" ) if not self.keep_accents: _a : Optional[Any] = unicodedata.normalize("""NFKD""" , UpperCAmelCase__ ) _a : Dict = """""".join([c for c in outputs if not unicodedata.combining(UpperCAmelCase__ )] ) if self.do_lower_case: _a : Union[str, Any] = outputs.lower() return outputs def _lowercase ( self : List[str] , UpperCAmelCase__ : str ) -> List[str]: _a : str = self.preprocess_text(UpperCAmelCase__ ) _a : Dict = self.sp_model.encode(UpperCAmelCase__ , out_type=UpperCAmelCase__ ) _a : Union[str, Any] = [] for piece in pieces: if len(UpperCAmelCase__ ) > 1 and piece[-1] == str(""",""" ) and piece[-2].isdigit(): _a : Dict = self.sp_model.EncodeAsPieces(piece[:-1].replace(UpperCAmelCase__ , """""" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: _a : Dict = cur_pieces[1:] else: _a : Any = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(UpperCAmelCase__ ) else: new_pieces.append(UpperCAmelCase__ ) return new_pieces def _lowercase ( self : Union[str, Any] , UpperCAmelCase__ : int ) -> int: return self.sp_model.PieceToId(UpperCAmelCase__ ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : Optional[Any] ) -> Any: return self.sp_model.IdToPiece(UpperCAmelCase__ ) def _lowercase ( self : Any , UpperCAmelCase__ : Any ) -> Dict: _a : Dict = """""".join(UpperCAmelCase__ ).replace(UpperCAmelCase__ , """ """ ).strip() return out_string def _lowercase ( self : Any , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ) -> List[int]: _a : Optional[Any] = [self.sep_token_id] _a : Dict = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _lowercase ( self : Tuple , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None , UpperCAmelCase__ : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase__ , token_ids_a=UpperCAmelCase__ , already_has_special_tokens=UpperCAmelCase__ ) if token_ids_a is not None: return ([0] * len(UpperCAmelCase__ )) + [1] + ([0] * len(UpperCAmelCase__ )) + [1, 1] return ([0] * len(UpperCAmelCase__ )) + [1, 1] def _lowercase ( self : str , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ) -> List[int]: _a : Any = [self.sep_token_id] _a : Optional[Any] = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _lowercase ( self : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(UpperCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return _a : Union[str, Any] = os.path.join( UpperCAmelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCAmelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCAmelCase__ , """wb""" ) as fi: _a : Optional[Any] = self.sp_model.serialized_model_proto() fi.write(UpperCAmelCase__ ) return (out_vocab_file,) def _lowercase ( self : Any , *UpperCAmelCase__ : List[str] , **UpperCAmelCase__ : List[str] ) -> List[str]: _a : Tuple = super()._decode(*UpperCAmelCase__ , **UpperCAmelCase__ ) _a : Optional[Any] = text.replace(""" """ , """""" ).replace("""\u2582""" , """ """ ).replace("""\u2583""" , """\n""" ) return text
294
1
"""simple docstring""" import unittest from transformers import DebertaVaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaVaForMaskedLM, DebertaVaForMultipleChoice, DebertaVaForQuestionAnswering, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaModel, ) from transformers.models.deberta_va.modeling_deberta_va import DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCamelCase ( snake_case_ ): def __init__( self : List[str] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Tuple=13 , UpperCAmelCase__ : Any=7 , UpperCAmelCase__ : List[Any]=True , UpperCAmelCase__ : Optional[Any]=True , UpperCAmelCase__ : Dict=True , UpperCAmelCase__ : Dict=True , UpperCAmelCase__ : int=99 , UpperCAmelCase__ : Optional[Any]=32 , UpperCAmelCase__ : int=5 , UpperCAmelCase__ : int=4 , UpperCAmelCase__ : Dict=37 , UpperCAmelCase__ : Optional[Any]="gelu" , UpperCAmelCase__ : int=0.1 , UpperCAmelCase__ : Union[str, Any]=0.1 , UpperCAmelCase__ : List[str]=512 , UpperCAmelCase__ : Tuple=16 , UpperCAmelCase__ : Optional[Any]=2 , UpperCAmelCase__ : Optional[Any]=0.0_2 , UpperCAmelCase__ : Any=False , UpperCAmelCase__ : Tuple=True , UpperCAmelCase__ : Any="None" , UpperCAmelCase__ : Optional[int]=3 , UpperCAmelCase__ : List[Any]=4 , UpperCAmelCase__ : Optional[int]=None , ) -> List[str]: _a : Any = parent _a : Union[str, Any] = batch_size _a : List[str] = seq_length _a : List[str] = is_training _a : Optional[Any] = use_input_mask _a : Any = use_token_type_ids _a : str = use_labels _a : Tuple = vocab_size _a : List[str] = hidden_size _a : List[str] = num_hidden_layers _a : List[Any] = num_attention_heads _a : Tuple = intermediate_size _a : Dict = hidden_act _a : Tuple = hidden_dropout_prob _a : Tuple = attention_probs_dropout_prob _a : Dict = max_position_embeddings _a : Tuple = type_vocab_size _a : Optional[Any] = type_sequence_label_size _a : int = initializer_range _a : Optional[int] = num_labels _a : List[str] = num_choices _a : Any = relative_attention _a : Optional[Any] = position_biased_input _a : List[Any] = pos_att_type _a : str = scope def _lowercase ( self : int ) -> Any: _a : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _a : Optional[int] = None if self.use_input_mask: _a : str = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) _a : Optional[int] = None if self.use_token_type_ids: _a : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _a : Union[str, Any] = None _a : Optional[Any] = None _a : int = None if self.use_labels: _a : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _a : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _a : List[str] = ids_tensor([self.batch_size] , self.num_choices ) _a : Optional[int] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowercase ( self : List[str] ) -> Optional[Any]: return DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : Union[str, Any] ) -> Union[str, Any]: self.parent.assertListEqual(list(result.loss.size() ) , [] ) def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : str , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Any , UpperCAmelCase__ : Union[str, Any] ) -> Tuple: _a : Any = DebertaVaModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() _a : Union[str, Any] = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ )[0] _a : List[str] = model(UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ )[0] _a : List[str] = model(UpperCAmelCase__ )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def _lowercase ( self : Dict , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Any ) -> int: _a : Optional[Any] = DebertaVaForMaskedLM(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() _a : List[str] = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : str ) -> str: _a : List[str] = self.num_labels _a : List[Any] = DebertaVaForSequenceClassification(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() _a : Dict = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(UpperCAmelCase__ ) def _lowercase ( self : List[str] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : str , UpperCAmelCase__ : str , UpperCAmelCase__ : int ) -> Any: _a : Optional[Any] = self.num_labels _a : Any = DebertaVaForTokenClassification(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() _a : Union[str, Any] = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowercase ( self : str , UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Dict ) -> Optional[Any]: _a : int = DebertaVaForQuestionAnswering(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() _a : str = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , start_positions=UpperCAmelCase__ , end_positions=UpperCAmelCase__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowercase ( self : List[str] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : int , UpperCAmelCase__ : Any , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : List[str] ) -> Union[str, Any]: _a : List[Any] = DebertaVaForMultipleChoice(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() _a : Optional[int] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _a : Optional[Any] = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _a : Any = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _a : Any = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _lowercase ( self : int ) -> Dict: _a : int = self.prepare_config_and_inputs() ( ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ) : Any = config_and_inputs _a : Tuple = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class UpperCamelCase ( snake_case_ , snake_case_ , unittest.TestCase ): UpperCamelCase : Optional[int] = ( ( DebertaVaModel, DebertaVaForMaskedLM, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaForQuestionAnswering, DebertaVaForMultipleChoice, ) if is_torch_available() else () ) UpperCamelCase : Optional[int] = ( { '''feature-extraction''': DebertaVaModel, '''fill-mask''': DebertaVaForMaskedLM, '''question-answering''': DebertaVaForQuestionAnswering, '''text-classification''': DebertaVaForSequenceClassification, '''token-classification''': DebertaVaForTokenClassification, '''zero-shot''': DebertaVaForSequenceClassification, } if is_torch_available() else {} ) UpperCamelCase : Tuple = True UpperCamelCase : Any = False UpperCamelCase : Tuple = False UpperCamelCase : int = False UpperCamelCase : str = False def _lowercase ( self : str ) -> List[Any]: _a : List[str] = DebertaVaModelTester(self ) _a : int = ConfigTester(self , config_class=UpperCAmelCase__ , hidden_size=37 ) def _lowercase ( self : str ) -> Dict: self.config_tester.run_common_tests() def _lowercase ( self : Dict ) -> List[str]: _a : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*UpperCAmelCase__ ) def _lowercase ( self : Tuple ) -> str: _a : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*UpperCAmelCase__ ) def _lowercase ( self : Union[str, Any] ) -> List[Any]: _a : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*UpperCAmelCase__ ) def _lowercase ( self : Optional[int] ) -> int: _a : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*UpperCAmelCase__ ) def _lowercase ( self : str ) -> Union[str, Any]: _a : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*UpperCAmelCase__ ) def _lowercase ( self : Any ) -> Tuple: _a : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_multiple_choice(*UpperCAmelCase__ ) @slow def _lowercase ( self : Tuple ) -> Optional[int]: for model_name in DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _a : Dict = DebertaVaModel.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) @require_torch @require_sentencepiece @require_tokenizers class UpperCamelCase ( unittest.TestCase ): @unittest.skip(reason="""Model not available yet""" ) def _lowercase ( self : Union[str, Any] ) -> Optional[Any]: pass @slow def _lowercase ( self : List[Any] ) -> Union[str, Any]: _a : str = DebertaVaModel.from_pretrained("""microsoft/deberta-v2-xlarge""" ) _a : str = torch.tensor([[0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2]] ) _a : str = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): _a : List[Any] = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ )[0] # compare the actual values for a slice. _a : List[Any] = torch.tensor( [[[0.2_3_5_6, 0.1_9_4_8, 0.0_3_6_9], [-0.1_0_6_3, 0.3_5_8_6, -0.5_1_5_2], [-0.6_3_9_9, -0.0_2_5_9, -0.2_5_2_5]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , UpperCAmelCase__ , atol=1E-4 ) , f"""{output[:, 1:4, 1:4]}""" )
294
"""simple docstring""" import inspect import unittest import torch import torch.nn as nn from accelerate.hooks import ( AlignDevicesHook, ModelHook, SequentialHook, add_hook_to_module, attach_align_device_hook, remove_hook_from_module, remove_hook_from_submodules, ) from accelerate.test_utils import require_multi_gpu class UpperCamelCase ( nn.Module ): def __init__( self : Union[str, Any] ) -> int: super().__init__() _a : Optional[Any] = nn.Linear(3 , 4 ) _a : Tuple = nn.BatchNormad(4 ) _a : Dict = nn.Linear(4 , 5 ) def _lowercase ( self : Optional[int] , UpperCAmelCase__ : List[str] ) -> int: return self.lineara(self.batchnorm(self.lineara(UpperCAmelCase__ ) ) ) class UpperCamelCase ( snake_case_ ): def _lowercase ( self : Any , UpperCAmelCase__ : Any , *UpperCAmelCase__ : List[str] , **UpperCAmelCase__ : Optional[int] ) -> Optional[int]: return (args[0] + 1,) + args[1:], kwargs class UpperCamelCase ( snake_case_ ): def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Optional[Any] ) -> List[str]: return output + 1 class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : Dict ) -> str: _a : List[Any] = ModelForTest() _a : str = ModelHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) self.assertEqual(test_model._hf_hook , UpperCAmelCase__ ) self.assertTrue(hasattr(UpperCAmelCase__ , """_old_forward""" ) ) # Check adding the hook did not change the name or the signature self.assertEqual(test_model.forward.__name__ , """forward""" ) self.assertListEqual(list(inspect.signature(test_model.forward ).parameters ) , ["""x"""] ) remove_hook_from_module(UpperCAmelCase__ ) self.assertFalse(hasattr(UpperCAmelCase__ , """_hf_hook""" ) ) self.assertFalse(hasattr(UpperCAmelCase__ , """_old_forward""" ) ) def _lowercase ( self : Optional[int] ) -> Optional[int]: _a : Dict = ModelForTest() _a : Dict = ModelHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ , append=UpperCAmelCase__ ) self.assertEqual(isinstance(test_model._hf_hook , UpperCAmelCase__ ) , UpperCAmelCase__ ) self.assertEqual(len(test_model._hf_hook.hooks ) , 2 ) self.assertTrue(hasattr(UpperCAmelCase__ , """_old_forward""" ) ) # Check adding the hook did not change the name or the signature self.assertEqual(test_model.forward.__name__ , """forward""" ) self.assertListEqual(list(inspect.signature(test_model.forward ).parameters ) , ["""x"""] ) remove_hook_from_module(UpperCAmelCase__ ) self.assertFalse(hasattr(UpperCAmelCase__ , """_hf_hook""" ) ) self.assertFalse(hasattr(UpperCAmelCase__ , """_old_forward""" ) ) def _lowercase ( self : Dict ) -> int: _a : str = ModelForTest() _a : List[Any] = torch.randn(2 , 3 ) _a : Optional[Any] = test_model(x + 1 ) _a : str = test_model(x + 2 ) _a : Union[str, Any] = PreForwardHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Tuple = test_model(UpperCAmelCase__ ) self.assertTrue(torch.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1E-5 ) ) # Attaching a hook to a model when it already has one replaces, does not chain _a : int = PreForwardHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : str = test_model(UpperCAmelCase__ ) self.assertTrue(torch.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1E-5 ) ) # You need to use the sequential hook to chain two or more hooks _a : int = SequentialHook(PreForwardHook() , PreForwardHook() ) add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Tuple = test_model(UpperCAmelCase__ ) assert torch.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1E-5 ) def _lowercase ( self : Tuple ) -> int: _a : Tuple = ModelForTest() _a : Union[str, Any] = torch.randn(2 , 3 ) _a : Optional[int] = test_model(UpperCAmelCase__ ) _a : int = PostForwardHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[str] = test_model(UpperCAmelCase__ ) self.assertTrue(torch.allclose(UpperCAmelCase__ , output + 1 , atol=1E-5 ) ) # Attaching a hook to a model when it already has one replaces, does not chain _a : List[Any] = PostForwardHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Dict = test_model(UpperCAmelCase__ ) self.assertTrue(torch.allclose(UpperCAmelCase__ , output + 1 , atol=1E-5 ) ) # You need to use the sequential hook to chain two or more hooks _a : Any = SequentialHook(PostForwardHook() , PostForwardHook() ) add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Optional[int] = test_model(UpperCAmelCase__ ) assert torch.allclose(UpperCAmelCase__ , output + 2 , atol=1E-5 ) def _lowercase ( self : Dict ) -> Optional[Any]: _a : Any = ModelForTest() _a : List[Any] = torch.randn(2 , 3 ) _a : Dict = test_model(UpperCAmelCase__ ) _a : Any = PostForwardHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[str] = test_model(UpperCAmelCase__ ) self.assertTrue(torch.allclose(UpperCAmelCase__ , output + 1 ) ) self.assertTrue(outputa.requires_grad ) _a : Any = True _a : Union[str, Any] = test_model(UpperCAmelCase__ ) self.assertFalse(outputa.requires_grad ) @require_multi_gpu def _lowercase ( self : Optional[Any] ) -> str: _a : List[Any] = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices add_hook_to_module(model.lineara , AlignDevicesHook(execution_device=0 ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(execution_device=0 ) ) add_hook_to_module(model.lineara , AlignDevicesHook(execution_device=1 ) ) self.assertEqual(model.lineara.weight.device , torch.device(0 ) ) self.assertEqual(model.batchnorm.weight.device , torch.device(0 ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device(0 ) ) self.assertEqual(model.lineara.weight.device , torch.device(1 ) ) # We can still make a forward pass. The input does not need to be on any particular device _a : Optional[int] = torch.randn(2 , 3 ) _a : Any = model(UpperCAmelCase__ ) self.assertEqual(output.device , torch.device(1 ) ) # We can add a general hook to put back output on same device as input. add_hook_to_module(UpperCAmelCase__ , AlignDevicesHook(io_same_device=UpperCAmelCase__ ) ) _a : str = torch.randn(2 , 3 ).to(0 ) _a : Union[str, Any] = model(UpperCAmelCase__ ) self.assertEqual(output.device , torch.device(0 ) ) def _lowercase ( self : str ) -> Union[str, Any]: _a : int = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices _a : List[Any] = {"""execution_device""": 0 if torch.cuda.is_available() else """cpu""", """offload""": True} add_hook_to_module(model.lineara , AlignDevicesHook(**UpperCAmelCase__ ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(**UpperCAmelCase__ ) ) add_hook_to_module(model.lineara , AlignDevicesHook(**UpperCAmelCase__ ) ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) # Buffers are not included in the offload by default, so are on the execution device _a : Dict = torch.device(hook_kwargs["""execution_device"""] ) self.assertEqual(model.batchnorm.running_mean.device , UpperCAmelCase__ ) _a : int = torch.randn(2 , 3 ) _a : str = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_module(model.lineara ) remove_hook_from_module(model.batchnorm ) remove_hook_from_module(model.lineara ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # Now test with buffers included in the offload _a : List[str] = { """execution_device""": 0 if torch.cuda.is_available() else """cpu""", """offload""": True, """offload_buffers""": True, } add_hook_to_module(model.lineara , AlignDevicesHook(**UpperCAmelCase__ ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(**UpperCAmelCase__ ) ) add_hook_to_module(model.lineara , AlignDevicesHook(**UpperCAmelCase__ ) ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device("""meta""" ) ) _a : Tuple = torch.randn(2 , 3 ) _a : Union[str, Any] = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_module(model.lineara ) remove_hook_from_module(model.batchnorm ) remove_hook_from_module(model.lineara ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) def _lowercase ( self : Tuple ) -> List[str]: _a : str = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices _a : Union[str, Any] = 0 if torch.cuda.is_available() else """cpu""" attach_align_device_hook(UpperCAmelCase__ , execution_device=UpperCAmelCase__ , offload=UpperCAmelCase__ ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) # Buffers are not included in the offload by default, so are on the execution device _a : Dict = torch.device(UpperCAmelCase__ ) self.assertEqual(model.batchnorm.running_mean.device , UpperCAmelCase__ ) _a : Union[str, Any] = torch.randn(2 , 3 ) _a : List[Any] = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(UpperCAmelCase__ ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # Now test with buffers included in the offload attach_align_device_hook(UpperCAmelCase__ , execution_device=UpperCAmelCase__ , offload=UpperCAmelCase__ , offload_buffers=UpperCAmelCase__ ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device("""meta""" ) ) _a : List[str] = torch.randn(2 , 3 ) _a : Union[str, Any] = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(UpperCAmelCase__ ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) def _lowercase ( self : Dict ) -> str: _a : Optional[Any] = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices _a : str = 0 if torch.cuda.is_available() else """cpu""" attach_align_device_hook( UpperCAmelCase__ , execution_device=UpperCAmelCase__ , offload=UpperCAmelCase__ , weights_map=model.state_dict() ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) # Buffers are not included in the offload by default, so are on the execution device _a : Union[str, Any] = torch.device(UpperCAmelCase__ ) self.assertEqual(model.batchnorm.running_mean.device , UpperCAmelCase__ ) _a : Union[str, Any] = torch.randn(2 , 3 ) _a : int = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(UpperCAmelCase__ ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # Now test with buffers included in the offload attach_align_device_hook( UpperCAmelCase__ , execution_device=UpperCAmelCase__ , offload=UpperCAmelCase__ , weights_map=model.state_dict() , offload_buffers=UpperCAmelCase__ , ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device("""meta""" ) ) _a : Any = torch.randn(2 , 3 ) _a : int = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(UpperCAmelCase__ ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) )
294
1
"""simple docstring""" def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' if not isinstance(UpperCamelCase__ , UpperCamelCase__ ) or number < 0: raise ValueError("""Input must be a non-negative integer""" ) _a : Optional[Any] = 0 while number: # This way we arrive at next set bit (next 1) instead of looping # through each bit and checking for 1s hence the # loop won't run 32 times it will only run the number of `1` times number &= number - 1 count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
294
"""simple docstring""" from __future__ import annotations def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' print(F"""Vertex\tShortest Distance from vertex {src}""" ) for i, d in enumerate(UpperCamelCase__ ): print(F"""{i}\t\t{d}""" ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' for j in range(UpperCamelCase__ ): _a , _a , _a : List[str] = (graph[j][k] for k in ["""src""", """dst""", """weight"""]) if distance[u] != float("""inf""" ) and distance[u] + w < distance[v]: return True return False def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Dict = [float("""inf""" )] * vertex_count _a : Any = 0.0 for _ in range(vertex_count - 1 ): for j in range(UpperCamelCase__ ): _a , _a , _a : List[Any] = (graph[j][k] for k in ["""src""", """dst""", """weight"""]) if distance[u] != float("""inf""" ) and distance[u] + w < distance[v]: _a : Any = distance[u] + w _a : Union[str, Any] = check_negative_cycle(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) if negative_cycle_exists: raise Exception("""Negative cycle found""" ) return distance if __name__ == "__main__": import doctest doctest.testmod() _snake_case = int(input('Enter number of vertices: ').strip()) _snake_case = int(input('Enter number of edges: ').strip()) _snake_case = [{} for _ in range(E)] for i in range(E): print('Edge ', i + 1) _snake_case , _snake_case , _snake_case = ( int(x) for x in input('Enter source, destination, weight: ').strip().split(' ') ) _snake_case = {'src': src, 'dst': dest, 'weight': weight} _snake_case = int(input('\nEnter shortest path source:').strip()) _snake_case = bellman_ford(graph, V, E, source) print_distance(shortest_distance, 0)
294
1
"""simple docstring""" import logging import os import threading import time try: import warnings except ImportError: _snake_case = None try: import msvcrt except ImportError: _snake_case = None try: import fcntl except ImportError: _snake_case = None # Backward compatibility # ------------------------------------------------ try: TimeoutError except NameError: _snake_case = OSError # Data # ------------------------------------------------ _snake_case = [ 'Timeout', 'BaseFileLock', 'WindowsFileLock', 'UnixFileLock', 'SoftFileLock', 'FileLock', ] _snake_case = '3.0.12' _snake_case = None def lowerCAmelCase__ ( ): '''simple docstring''' global _logger _a : List[str] = _logger or logging.getLogger(__name__ ) return _logger class UpperCamelCase ( snake_case_ ): def __init__( self : int , UpperCAmelCase__ : str ) -> Optional[int]: _a : Optional[Any] = lock_file return None def __str__( self : str ) -> Dict: _a : Any = f"""The file lock '{self.lock_file}' could not be acquired.""" return temp class UpperCamelCase : def __init__( self : Optional[int] , UpperCAmelCase__ : Optional[int] ) -> List[Any]: _a : Dict = lock return None def __enter__( self : List[Any] ) -> int: return self.lock def __exit__( self : List[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[int] ) -> List[str]: self.lock.release() return None class UpperCamelCase : def __init__( self : Dict , UpperCAmelCase__ : str , UpperCAmelCase__ : Union[str, Any]=-1 , UpperCAmelCase__ : List[Any]=None ) -> Any: _a : Optional[Any] = max_filename_length if max_filename_length is not None else 255 # Hash the filename if it's too long _a : Tuple = self.hash_filename_if_too_long(UpperCAmelCase__ , UpperCAmelCase__ ) # The path to the lock file. _a : Dict = lock_file # The file descriptor for the *_lock_file* as it is returned by the # os.open() function. # This file lock is only NOT None, if the object currently holds the # lock. _a : str = None # The default timeout value. _a : Tuple = timeout # We use this lock primarily for the lock counter. _a : Any = threading.Lock() # The lock counter is used for implementing the nested locking # mechanism. Whenever the lock is acquired, the counter is increased and # the lock is only released, when this value is 0 again. _a : Optional[int] = 0 return None @property def _lowercase ( self : List[str] ) -> Union[str, Any]: return self._lock_file @property def _lowercase ( self : Optional[Any] ) -> Dict: return self._timeout @timeout.setter def _lowercase ( self : Union[str, Any] , UpperCAmelCase__ : Tuple ) -> Optional[Any]: _a : int = float(UpperCAmelCase__ ) return None def _lowercase ( self : List[str] ) -> List[Any]: raise NotImplementedError() def _lowercase ( self : int ) -> Union[str, Any]: raise NotImplementedError() @property def _lowercase ( self : Optional[Any] ) -> Dict: return self._lock_file_fd is not None def _lowercase ( self : List[Any] , UpperCAmelCase__ : Optional[Any]=None , UpperCAmelCase__ : Optional[Any]=0.0_5 ) -> Union[str, Any]: # Use the default timeout, if no timeout is provided. if timeout is None: _a : Dict = self.timeout # Increment the number right at the beginning. # We can still undo it, if something fails. with self._thread_lock: self._lock_counter += 1 _a : Optional[int] = id(self ) _a : Any = self._lock_file _a : Optional[int] = time.time() try: while True: with self._thread_lock: if not self.is_locked: logger().debug(f"""Attempting to acquire lock {lock_id} on {lock_filename}""" ) self._acquire() if self.is_locked: logger().debug(f"""Lock {lock_id} acquired on {lock_filename}""" ) break elif timeout >= 0 and time.time() - start_time > timeout: logger().debug(f"""Timeout on acquiring lock {lock_id} on {lock_filename}""" ) raise Timeout(self._lock_file ) else: logger().debug( f"""Lock {lock_id} not acquired on {lock_filename}, waiting {poll_intervall} seconds ...""" ) time.sleep(UpperCAmelCase__ ) except: # noqa # Something did go wrong, so decrement the counter. with self._thread_lock: _a : List[Any] = max(0 , self._lock_counter - 1 ) raise return _Acquire_ReturnProxy(lock=self ) def _lowercase ( self : Tuple , UpperCAmelCase__ : Optional[Any]=False ) -> List[Any]: with self._thread_lock: if self.is_locked: self._lock_counter -= 1 if self._lock_counter == 0 or force: _a : str = id(self ) _a : Dict = self._lock_file logger().debug(f"""Attempting to release lock {lock_id} on {lock_filename}""" ) self._release() _a : int = 0 logger().debug(f"""Lock {lock_id} released on {lock_filename}""" ) return None def __enter__( self : Optional[int] ) -> str: self.acquire() return self def __exit__( self : Optional[int] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Tuple ) -> Optional[int]: self.release() return None def __del__( self : Tuple ) -> Optional[int]: self.release(force=UpperCAmelCase__ ) return None def _lowercase ( self : Tuple , UpperCAmelCase__ : str , UpperCAmelCase__ : int ) -> str: _a : Union[str, Any] = os.path.basename(UpperCAmelCase__ ) if len(UpperCAmelCase__ ) > max_length and max_length > 0: _a : List[Any] = os.path.dirname(UpperCAmelCase__ ) _a : Union[str, Any] = str(hash(UpperCAmelCase__ ) ) _a : Tuple = filename[: max_length - len(UpperCAmelCase__ ) - 8] + """...""" + hashed_filename + """.lock""" return os.path.join(UpperCAmelCase__ , UpperCAmelCase__ ) else: return path class UpperCamelCase ( snake_case_ ): def __init__( self : Any , UpperCAmelCase__ : int , UpperCAmelCase__ : Tuple=-1 , UpperCAmelCase__ : Any=None ) -> Any: from .file_utils import relative_to_absolute_path super().__init__(UpperCAmelCase__ , timeout=UpperCAmelCase__ , max_filename_length=UpperCAmelCase__ ) _a : Dict = """\\\\?\\""" + relative_to_absolute_path(self.lock_file ) def _lowercase ( self : Optional[int] ) -> List[Any]: _a : Tuple = os.O_RDWR | os.O_CREAT | os.O_TRUNC try: _a : List[Any] = os.open(self._lock_file , UpperCAmelCase__ ) except OSError: pass else: try: msvcrt.locking(UpperCAmelCase__ , msvcrt.LK_NBLCK , 1 ) except OSError: os.close(UpperCAmelCase__ ) else: _a : List[Any] = fd return None def _lowercase ( self : List[Any] ) -> int: _a : str = self._lock_file_fd _a : Any = None msvcrt.locking(UpperCAmelCase__ , msvcrt.LK_UNLCK , 1 ) os.close(UpperCAmelCase__ ) try: os.remove(self._lock_file ) # Probably another instance of the application # that acquired the file lock. except OSError: pass return None class UpperCamelCase ( snake_case_ ): def __init__( self : Tuple , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : List[str]=-1 , UpperCAmelCase__ : List[str]=None ) -> int: _a : Dict = os.statvfs(os.path.dirname(UpperCAmelCase__ ) ).f_namemax super().__init__(UpperCAmelCase__ , timeout=UpperCAmelCase__ , max_filename_length=UpperCAmelCase__ ) def _lowercase ( self : Tuple ) -> Any: _a : List[str] = os.O_RDWR | os.O_CREAT | os.O_TRUNC _a : Optional[Any] = os.open(self._lock_file , UpperCAmelCase__ ) try: fcntl.flock(UpperCAmelCase__ , fcntl.LOCK_EX | fcntl.LOCK_NB ) except OSError: os.close(UpperCAmelCase__ ) else: _a : Optional[int] = fd return None def _lowercase ( self : List[Any] ) -> Dict: # Do not remove the lockfile: # # https://github.com/benediktschmitt/py-filelock/issues/31 # https://stackoverflow.com/questions/17708885/flock-removing-locked-file-without-race-condition _a : List[str] = self._lock_file_fd _a : List[str] = None fcntl.flock(UpperCAmelCase__ , fcntl.LOCK_UN ) os.close(UpperCAmelCase__ ) return None class UpperCamelCase ( snake_case_ ): def _lowercase ( self : Tuple ) -> Any: _a : Tuple = os.O_WRONLY | os.O_CREAT | os.O_EXCL | os.O_TRUNC try: _a : Tuple = os.open(self._lock_file , UpperCAmelCase__ ) except OSError: pass else: _a : Union[str, Any] = fd return None def _lowercase ( self : Dict ) -> int: os.close(self._lock_file_fd ) _a : List[Any] = None try: os.remove(self._lock_file ) # The file is already deleted and that's what we want. except OSError: pass return None _snake_case = None if msvcrt: _snake_case = WindowsFileLock elif fcntl: _snake_case = UnixFileLock else: _snake_case = SoftFileLock if warnings is not None: warnings.warn('only soft file lock is available')
294
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _snake_case = { 'configuration_transfo_xl': ['TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TransfoXLConfig'], 'tokenization_transfo_xl': ['TransfoXLCorpus', 'TransfoXLTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST', 'AdaptiveEmbedding', 'TransfoXLForSequenceClassification', 'TransfoXLLMHeadModel', 'TransfoXLModel', 'TransfoXLPreTrainedModel', 'load_tf_weights_in_transfo_xl', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFAdaptiveEmbedding', 'TFTransfoXLForSequenceClassification', 'TFTransfoXLLMHeadModel', 'TFTransfoXLMainLayer', 'TFTransfoXLModel', 'TFTransfoXLPreTrainedModel', ] if TYPE_CHECKING: from .configuration_transfo_xl import TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, TransfoXLConfig from .tokenization_transfo_xl import TransfoXLCorpus, TransfoXLTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_transfo_xl import ( TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, AdaptiveEmbedding, TransfoXLForSequenceClassification, TransfoXLLMHeadModel, TransfoXLModel, TransfoXLPreTrainedModel, load_tf_weights_in_transfo_xl, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_transfo_xl import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFAdaptiveEmbedding, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLMainLayer, TFTransfoXLModel, TFTransfoXLPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
294
1
"""simple docstring""" from .imports import is_tqdm_available if is_tqdm_available(): from tqdm.auto import tqdm as _tqdm from ..state import PartialState def lowerCAmelCase__ ( UpperCamelCase__ = True , *UpperCamelCase__ , **UpperCamelCase__ ): '''simple docstring''' if not is_tqdm_available(): raise ImportError("""Accelerate's `tqdm` module requires `tqdm` to be installed. Please run `pip install tqdm`.""" ) _a : Optional[int] = False if main_process_only: _a : Tuple = PartialState().local_process_index == 0 return _tqdm(*UpperCamelCase__ , **UpperCamelCase__ , disable=UpperCamelCase__ )
294
"""simple docstring""" def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , ): '''simple docstring''' _a : Optional[Any] = [redshift, radiation_density, matter_density, dark_energy] if any(p < 0 for p in parameters ): raise ValueError("""All input parameters must be positive""" ) if any(p > 1 for p in parameters[1:4] ): raise ValueError("""Relative densities cannot be greater than one""" ) else: _a : Tuple = 1 - (matter_density + radiation_density + dark_energy) _a : int = ( radiation_density * (redshift + 1) ** 4 + matter_density * (redshift + 1) ** 3 + curvature * (redshift + 1) ** 2 + dark_energy ) _a : List[str] = hubble_constant * e_a ** (1 / 2) return hubble if __name__ == "__main__": import doctest # run doctest doctest.testmod() # demo LCDM approximation _snake_case = 0.3 print( hubble_parameter( hubble_constant=68.3, radiation_density=1e-4, matter_density=matter_density, dark_energy=1 - matter_density, redshift=0, ) )
294
1
"""simple docstring""" from math import factorial _snake_case = {str(digit): factorial(digit) for digit in range(10)} def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise TypeError("""Parameter number must be int""" ) if number < 0: raise ValueError("""Parameter number must be greater than or equal to 0""" ) # Converts number in string to iterate on its digits and adds its factorial. return sum(DIGIT_FACTORIAL[digit] for digit in str(UpperCamelCase__ ) ) def lowerCAmelCase__ ( UpperCamelCase__ = 6_0 , UpperCamelCase__ = 1_0_0_0_0_0_0 ): '''simple docstring''' if not isinstance(UpperCamelCase__ , UpperCamelCase__ ) or not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise TypeError("""Parameters chain_length and number_limit must be int""" ) if chain_length <= 0 or number_limit <= 0: raise ValueError( """Parameters chain_length and number_limit must be greater than 0""" ) # the counter for the chains with the exact desired length _a : List[str] = 0 # the cached sizes of the previous chains _a : dict[int, int] = {} for start_chain_element in range(1 , UpperCamelCase__ ): # The temporary set will contain the elements of the chain _a : Any = set() _a : int = 0 # Stop computing the chain when you find a cached size, a repeating item or the # length is greater then the desired one. _a : Tuple = start_chain_element while ( chain_element not in chain_sets_lengths and chain_element not in chain_set and chain_set_length <= chain_length ): chain_set.add(UpperCamelCase__ ) chain_set_length += 1 _a : Optional[Any] = digit_factorial_sum(UpperCamelCase__ ) if chain_element in chain_sets_lengths: chain_set_length += chain_sets_lengths[chain_element] _a : List[str] = chain_set_length # If chain contains the exact amount of elements increase the counter if chain_set_length == chain_length: chains_counter += 1 return chains_counter if __name__ == "__main__": import doctest doctest.testmod() print(F'''{solution()}''')
294
"""simple docstring""" import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : List[Any] ) -> Dict: _a : Optional[int] = tempfile.mkdtemp() _a : Optional[Any] = SamImageProcessor() _a : int = SamProcessor(UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) def _lowercase ( self : Tuple , **UpperCAmelCase__ : Any ) -> Any: return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase__ ).image_processor def _lowercase ( self : str ) -> int: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Tuple ) -> Dict: _a : List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _a : Optional[int] = [Image.fromarray(np.moveaxis(UpperCAmelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def _lowercase ( self : Dict ) -> Dict: _a : List[Any] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _a : Tuple = self.get_image_processor(do_normalize=UpperCAmelCase__ , padding_value=1.0 ) _a : Tuple = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=UpperCAmelCase__ , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCAmelCase__ ) def _lowercase ( self : Union[str, Any] ) -> Tuple: _a : Optional[Any] = self.get_image_processor() _a : int = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Union[str, Any] = self.prepare_image_inputs() _a : List[str] = image_processor(UpperCAmelCase__ , return_tensors="""np""" ) _a : List[str] = processor(images=UpperCAmelCase__ , return_tensors="""np""" ) input_feat_extract.pop("""original_sizes""" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("""reshaped_input_sizes""" ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_torch def _lowercase ( self : Optional[Any] ) -> Optional[Any]: _a : Optional[Any] = self.get_image_processor() _a : Dict = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Tuple = [torch.ones((1, 3, 5, 5) )] _a : Tuple = [[1764, 2646]] _a : Optional[int] = [[683, 1024]] _a : List[Any] = processor.post_process_masks(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) _a : int = processor.post_process_masks( UpperCAmelCase__ , torch.tensor(UpperCAmelCase__ ) , torch.tensor(UpperCAmelCase__ ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np _a : Optional[Any] = [np.ones((1, 3, 5, 5) )] _a : Tuple = processor.post_process_masks(UpperCAmelCase__ , np.array(UpperCAmelCase__ ) , np.array(UpperCAmelCase__ ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) _a : List[str] = [[1, 0], [0, 1]] with self.assertRaises(UpperCAmelCase__ ): _a : str = processor.post_process_masks(UpperCAmelCase__ , np.array(UpperCAmelCase__ ) , np.array(UpperCAmelCase__ ) ) @require_vision @require_tf class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : Any ) -> List[str]: _a : List[str] = tempfile.mkdtemp() _a : Any = SamImageProcessor() _a : Union[str, Any] = SamProcessor(UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) def _lowercase ( self : List[str] , **UpperCAmelCase__ : Any ) -> List[str]: return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase__ ).image_processor def _lowercase ( self : Optional[Any] ) -> Union[str, Any]: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Dict ) -> List[str]: _a : List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _a : List[Any] = [Image.fromarray(np.moveaxis(UpperCAmelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def _lowercase ( self : Union[str, Any] ) -> Union[str, Any]: _a : Optional[int] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _a : str = self.get_image_processor(do_normalize=UpperCAmelCase__ , padding_value=1.0 ) _a : Union[str, Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=UpperCAmelCase__ , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCAmelCase__ ) def _lowercase ( self : List[Any] ) -> str: _a : Union[str, Any] = self.get_image_processor() _a : Dict = SamProcessor(image_processor=UpperCAmelCase__ ) _a : int = self.prepare_image_inputs() _a : List[str] = image_processor(UpperCAmelCase__ , return_tensors="""np""" ) _a : List[str] = processor(images=UpperCAmelCase__ , return_tensors="""np""" ) input_feat_extract.pop("""original_sizes""" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("""reshaped_input_sizes""" ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_tf def _lowercase ( self : Optional[Any] ) -> int: _a : Optional[Any] = self.get_image_processor() _a : Dict = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Any = [tf.ones((1, 3, 5, 5) )] _a : Tuple = [[1764, 2646]] _a : str = [[683, 1024]] _a : Union[str, Any] = processor.post_process_masks(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , return_tensors="""tf""" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) _a : Union[str, Any] = processor.post_process_masks( UpperCAmelCase__ , tf.convert_to_tensor(UpperCAmelCase__ ) , tf.convert_to_tensor(UpperCAmelCase__ ) , return_tensors="""tf""" , ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np _a : List[Any] = [np.ones((1, 3, 5, 5) )] _a : Optional[int] = processor.post_process_masks( UpperCAmelCase__ , np.array(UpperCAmelCase__ ) , np.array(UpperCAmelCase__ ) , return_tensors="""tf""" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) _a : Dict = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): _a : List[Any] = processor.post_process_masks( UpperCAmelCase__ , np.array(UpperCAmelCase__ ) , np.array(UpperCAmelCase__ ) , return_tensors="""tf""" ) @require_vision @require_torchvision class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : str ) -> Optional[Any]: _a : Optional[Any] = tempfile.mkdtemp() _a : Dict = SamImageProcessor() _a : List[str] = SamProcessor(UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) def _lowercase ( self : Any , **UpperCAmelCase__ : Dict ) -> int: return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase__ ).image_processor def _lowercase ( self : Tuple ) -> List[Any]: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : str ) -> int: _a : str = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _a : int = [Image.fromarray(np.moveaxis(UpperCAmelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def _lowercase ( self : int ) -> List[Any]: _a : Optional[Any] = self.get_image_processor() _a : Optional[Any] = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Tuple = np.random.randint(0 , 2 , size=(1, 3, 5, 5) ).astype(np.floataa ) _a : str = [tf.convert_to_tensor(UpperCAmelCase__ )] _a : Optional[int] = [torch.tensor(UpperCAmelCase__ )] _a : Union[str, Any] = [[1764, 2646]] _a : List[str] = [[683, 1024]] _a : Optional[int] = processor.post_process_masks( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , return_tensors="""tf""" ) _a : List[str] = processor.post_process_masks( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , return_tensors="""pt""" ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def _lowercase ( self : str ) -> Optional[Any]: _a : List[Any] = self.get_image_processor() _a : Any = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Dict = self.prepare_image_inputs() _a : List[str] = image_processor(UpperCAmelCase__ , return_tensors="""pt""" )["""pixel_values"""].numpy() _a : str = processor(images=UpperCAmelCase__ , return_tensors="""pt""" )["""pixel_values"""].numpy() _a : Optional[Any] = image_processor(UpperCAmelCase__ , return_tensors="""tf""" )["""pixel_values"""].numpy() _a : Optional[int] = processor(images=UpperCAmelCase__ , return_tensors="""tf""" )["""pixel_values"""].numpy() self.assertTrue(np.allclose(UpperCAmelCase__ , UpperCAmelCase__ ) ) self.assertTrue(np.allclose(UpperCAmelCase__ , UpperCAmelCase__ ) ) self.assertTrue(np.allclose(UpperCAmelCase__ , UpperCAmelCase__ ) )
294
1
"""simple docstring""" import argparse import os from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_task_guides.py _snake_case = 'src/transformers' _snake_case = 'docs/source/en/tasks' def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' with open(UpperCamelCase__ , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: _a : Optional[int] = f.readlines() # Find the start prompt. _a : Tuple = 0 while not lines[start_index].startswith(UpperCamelCase__ ): start_index += 1 start_index += 1 _a : Union[str, Any] = start_index while not lines[end_index].startswith(UpperCamelCase__ ): end_index += 1 end_index -= 1 while len(lines[start_index] ) <= 1: start_index += 1 while len(lines[end_index] ) <= 1: end_index -= 1 end_index += 1 return "".join(lines[start_index:end_index] ), start_index, end_index, lines # This is to make sure the transformers module imported is the one in the repo. _snake_case = direct_transformers_import(TRANSFORMERS_PATH) _snake_case = { 'asr.md': transformers_module.models.auto.modeling_auto.MODEL_FOR_CTC_MAPPING_NAMES, 'audio_classification.md': transformers_module.models.auto.modeling_auto.MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES, 'language_modeling.md': transformers_module.models.auto.modeling_auto.MODEL_FOR_CAUSAL_LM_MAPPING_NAMES, 'image_classification.md': transformers_module.models.auto.modeling_auto.MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES, 'masked_language_modeling.md': transformers_module.models.auto.modeling_auto.MODEL_FOR_MASKED_LM_MAPPING_NAMES, 'multiple_choice.md': transformers_module.models.auto.modeling_auto.MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES, 'object_detection.md': transformers_module.models.auto.modeling_auto.MODEL_FOR_OBJECT_DETECTION_MAPPING_NAMES, 'question_answering.md': transformers_module.models.auto.modeling_auto.MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES, 'semantic_segmentation.md': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEMANTIC_SEGMENTATION_MAPPING_NAMES, 'sequence_classification.md': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES, 'summarization.md': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES, 'token_classification.md': transformers_module.models.auto.modeling_auto.MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES, 'translation.md': transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES, 'video_classification.md': transformers_module.models.auto.modeling_auto.MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING_NAMES, 'document_question_answering.md': transformers_module.models.auto.modeling_auto.MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING_NAMES, 'monocular_depth_estimation.md': transformers_module.models.auto.modeling_auto.MODEL_FOR_DEPTH_ESTIMATION_MAPPING_NAMES, } # This list contains model types used in some task guides that are not in `CONFIG_MAPPING_NAMES` (therefore not in any # `MODEL_MAPPING_NAMES` or any `MODEL_FOR_XXX_MAPPING_NAMES`). _snake_case = { 'summarization.md': ('nllb',), 'translation.md': ('nllb',), } def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : Union[str, Any] = TASK_GUIDE_TO_MODELS[task_guide] _a : Any = SPECIAL_TASK_GUIDE_TO_MODEL_TYPES.get(UpperCamelCase__ , set() ) _a : str = { code: name for code, name in transformers_module.MODEL_NAMES_MAPPING.items() if (code in model_maping_names or code in special_model_types) } return ", ".join([F"""[{name}](../model_doc/{code})""" for code, name in model_names.items()] ) + "\n" def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__=False ): '''simple docstring''' _a , _a , _a , _a : Any = _find_text_in_file( filename=os.path.join(UpperCamelCase__ , UpperCamelCase__ ) , start_prompt="""<!--This tip is automatically generated by `make fix-copies`, do not fill manually!-->""" , end_prompt="""<!--End of the generated tip-->""" , ) _a : Optional[int] = get_model_list_for_task(UpperCamelCase__ ) if current_list != new_list: if overwrite: with open(os.path.join(UpperCamelCase__ , UpperCamelCase__ ) , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(lines[:start_index] + [new_list] + lines[end_index:] ) else: raise ValueError( F"""The list of models that can be used in the {task_guide} guide needs an update. Run `make fix-copies`""" """ to fix this.""" ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() parser.add_argument('--fix_and_overwrite', action='store_true', help='Whether to fix inconsistencies.') _snake_case = parser.parse_args() for task_guide in TASK_GUIDE_TO_MODELS.keys(): check_model_list_for_task(task_guide, args.fix_and_overwrite)
294
"""simple docstring""" import argparse import gc import json import os import re import torch from huggingface_hub import hf_hub_download from transformers import AutoModelForCausalLM, AutoTokenizer, PreTrainedTokenizerFast, RwkvConfig from transformers.modeling_utils import WEIGHTS_INDEX_NAME, shard_checkpoint _snake_case = { '169M': 12, '430M': 24, '1B5': 24, '3B': 32, '7B': 32, '14B': 40, } _snake_case = { '169M': 768, '430M': 1024, '1B5': 2048, '3B': 2560, '7B': 4096, '14B': 5120, } def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : int = list(state_dict.keys() ) for name in state_dict_keys: _a : str = state_dict.pop(UpperCamelCase__ ) # emb -> embedding if name.startswith("""emb.""" ): _a : Dict = name.replace("""emb.""" , """embeddings.""" ) # ln_0 -> pre_ln (only present at block 0) if name.startswith("""blocks.0.ln0""" ): _a : Dict = name.replace("""blocks.0.ln0""" , """blocks.0.pre_ln""" ) # att -> attention _a : Any = re.sub(R"""blocks\.(\d+)\.att""" , R"""blocks.\1.attention""" , UpperCamelCase__ ) # ffn -> feed_forward _a : int = re.sub(R"""blocks\.(\d+)\.ffn""" , R"""blocks.\1.feed_forward""" , UpperCamelCase__ ) # time_mix_k -> time_mix_key and reshape if name.endswith(""".time_mix_k""" ): _a : List[str] = name.replace(""".time_mix_k""" , """.time_mix_key""" ) # time_mix_v -> time_mix_value and reshape if name.endswith(""".time_mix_v""" ): _a : Tuple = name.replace(""".time_mix_v""" , """.time_mix_value""" ) # time_mix_r -> time_mix_key and reshape if name.endswith(""".time_mix_r""" ): _a : Dict = name.replace(""".time_mix_r""" , """.time_mix_receptance""" ) if name != "head.weight": _a : Optional[int] = """rwkv.""" + name _a : Any = weight return state_dict def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=False , UpperCamelCase__=None ): '''simple docstring''' # 1. If possible, build the tokenizer. if tokenizer_file is None: print("""No `--tokenizer_file` provided, we will use the default tokenizer.""" ) _a : Tuple = 5_0_2_7_7 _a : str = AutoTokenizer.from_pretrained("""EleutherAI/gpt-neox-20b""" ) else: _a : int = PreTrainedTokenizerFast(tokenizer_file=UpperCamelCase__ ) _a : int = len(UpperCamelCase__ ) tokenizer.save_pretrained(UpperCamelCase__ ) # 2. Build the config _a : Optional[Any] = list(NUM_HIDDEN_LAYERS_MAPPING.keys() ) if size is None: # Try to infer size from the checkpoint name for candidate in possible_sizes: if candidate in checkpoint_file: _a : Tuple = candidate break if size is None: raise ValueError("""Could not infer the size, please provide it with the `--size` argument.""" ) if size not in possible_sizes: raise ValueError(F"""`size` should be one of {possible_sizes}, got {size}.""" ) _a : List[Any] = RwkvConfig( vocab_size=UpperCamelCase__ , num_hidden_layers=NUM_HIDDEN_LAYERS_MAPPING[size] , hidden_size=HIDEN_SIZE_MAPPING[size] , ) config.save_pretrained(UpperCamelCase__ ) # 3. Download model file then convert state_dict _a : str = hf_hub_download(UpperCamelCase__ , UpperCamelCase__ ) _a : int = torch.load(UpperCamelCase__ , map_location="""cpu""" ) _a : List[str] = convert_state_dict(UpperCamelCase__ ) # 4. Split in shards and save _a , _a : List[str] = shard_checkpoint(UpperCamelCase__ ) for shard_file, shard in shards.items(): torch.save(UpperCamelCase__ , os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) if index is not None: _a : Optional[Any] = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) # Save the index as well with open(UpperCamelCase__ , """w""" , encoding="""utf-8""" ) as f: _a : Dict = json.dumps(UpperCamelCase__ , indent=2 , sort_keys=UpperCamelCase__ ) + """\n""" f.write(UpperCamelCase__ ) # 5. Clean up shards (for some reason the file PyTorch saves take the same space as the whole state_dict print( """Cleaning up shards. This may error with an OOM error, it this is the case don't worry you still have converted the model.""" ) _a : List[Any] = list(shards.keys() ) del state_dict del shards gc.collect() for shard_file in shard_files: _a : Any = torch.load(os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) torch.save({k: v.cpu().clone() for k, v in state_dict.items()} , os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) del state_dict gc.collect() if push_to_hub: if model_name is None: raise ValueError("""Please provide a `model_name` to push the model to the Hub.""" ) _a : Dict = AutoModelForCausalLM.from_pretrained(UpperCamelCase__ ) model.push_to_hub(UpperCamelCase__ , max_shard_size="""2GB""" ) tokenizer.push_to_hub(UpperCamelCase__ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '--repo_id', default=None, type=str, required=True, help='Repo ID from which to pull the checkpoint.' ) parser.add_argument( '--checkpoint_file', default=None, type=str, required=True, help='Name of the checkpoint file in the repo.' ) parser.add_argument( '--output_dir', default=None, type=str, required=True, help='Where to save the converted model.' ) parser.add_argument( '--tokenizer_file', default=None, type=str, help='Path to the tokenizer file to use (if not provided, only the model is converted).', ) parser.add_argument( '--size', default=None, type=str, help='Size of the model. Will be inferred from the `checkpoint_file` if not passed.', ) parser.add_argument( '--push_to_hub', action='store_true', help='Push to the Hub the converted model.', ) parser.add_argument( '--model_name', default=None, type=str, help='Name of the pushed model on the Hub, including the username / organization.', ) _snake_case = parser.parse_args() convert_rmkv_checkpoint_to_hf_format( args.repo_id, args.checkpoint_file, args.output_dir, size=args.size, tokenizer_file=args.tokenizer_file, push_to_hub=args.push_to_hub, model_name=args.model_name, )
294
1
"""simple docstring""" import baseaa def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' return baseaa.aaaencode(string.encode("""utf-8""" ) ) def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' return baseaa.aaadecode(UpperCamelCase__ ).decode("""utf-8""" ) if __name__ == "__main__": import doctest doctest.testmod()
294
"""simple docstring""" import shutil import tempfile import unittest from transformers import ClapFeatureExtractor, ClapProcessor, RobertaTokenizer, RobertaTokenizerFast from transformers.testing_utils import require_sentencepiece, require_torchaudio from .test_feature_extraction_clap import floats_list @require_torchaudio @require_sentencepiece class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : int ) -> List[str]: _a : Any = """laion/clap-htsat-unfused""" _a : Union[str, Any] = tempfile.mkdtemp() def _lowercase ( self : List[Any] , **UpperCAmelCase__ : Any ) -> Dict: return RobertaTokenizer.from_pretrained(self.checkpoint , **UpperCAmelCase__ ) def _lowercase ( self : List[Any] , **UpperCAmelCase__ : List[str] ) -> int: return ClapFeatureExtractor.from_pretrained(self.checkpoint , **UpperCAmelCase__ ) def _lowercase ( self : List[Any] ) -> Tuple: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : List[str] ) -> Optional[int]: _a : List[str] = self.get_tokenizer() _a : Any = self.get_feature_extractor() _a : Optional[Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) _a : List[str] = ClapProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCAmelCase__ ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , UpperCAmelCase__ ) def _lowercase ( self : Dict ) -> Optional[int]: _a : Tuple = ClapProcessor(tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() ) processor.save_pretrained(self.tmpdirname ) _a : Dict = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) _a : Union[str, Any] = self.get_feature_extractor(do_normalize=UpperCAmelCase__ , padding_value=1.0 ) _a : Union[str, Any] = ClapProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=UpperCAmelCase__ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCAmelCase__ ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.feature_extractor , UpperCAmelCase__ ) def _lowercase ( self : List[str] ) -> Optional[Any]: _a : Optional[int] = self.get_feature_extractor() _a : Tuple = self.get_tokenizer() _a : List[Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) _a : Any = floats_list((3, 1000) ) _a : List[Any] = feature_extractor(UpperCAmelCase__ , return_tensors="""np""" ) _a : List[str] = processor(audios=UpperCAmelCase__ , return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def _lowercase ( self : Tuple ) -> Optional[int]: _a : List[str] = self.get_feature_extractor() _a : Any = self.get_tokenizer() _a : Any = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) _a : Optional[int] = """This is a test string""" _a : Tuple = processor(text=UpperCAmelCase__ ) _a : int = tokenizer(UpperCAmelCase__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _lowercase ( self : List[Any] ) -> Any: _a : str = self.get_feature_extractor() _a : List[str] = self.get_tokenizer() _a : List[Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) _a : Any = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _a : Dict = processor.batch_decode(UpperCAmelCase__ ) _a : Any = tokenizer.batch_decode(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def _lowercase ( self : Dict ) -> List[str]: _a : str = self.get_feature_extractor() _a : Optional[Any] = self.get_tokenizer() _a : Union[str, Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) self.assertListEqual( processor.model_input_names[2:] , feature_extractor.model_input_names , msg="""`processor` and `feature_extractor` model input names do not match""" , )
294
1
"""simple docstring""" from collections.abc import Iterable from typing import Generic, TypeVar _snake_case = TypeVar('_T') class UpperCamelCase ( Generic[_T] ): def __init__( self : Optional[int] , UpperCAmelCase__ : Iterable[_T] | None = None ) -> None: _a : list[_T] = list(iterable or [] ) _a : list[_T] = [] def __len__( self : str ) -> int: return len(self._stacka ) + len(self._stacka ) def __repr__( self : List[str] ) -> str: return f"""Queue({tuple(self._stacka[::-1] + self._stacka )})""" def _lowercase ( self : Union[str, Any] , UpperCAmelCase__ : _T ) -> None: self._stacka.append(UpperCAmelCase__ ) def _lowercase ( self : Optional[Any] ) -> _T: _a : Any = self._stacka.pop _a : Union[str, Any] = self._stacka.append if not self._stacka: while self._stacka: stacka_append(stacka_pop() ) if not self._stacka: raise IndexError("""Queue is empty""" ) return self._stacka.pop() if __name__ == "__main__": from doctest import testmod testmod()
294
"""simple docstring""" import warnings from ...utils import logging from .image_processing_clip import CLIPImageProcessor _snake_case = logging.get_logger(__name__) class UpperCamelCase ( snake_case_ ): def __init__( self : Any , *UpperCAmelCase__ : Optional[Any] , **UpperCAmelCase__ : Tuple ) -> None: warnings.warn( """The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use CLIPImageProcessor instead.""" , UpperCAmelCase__ , ) super().__init__(*UpperCAmelCase__ , **UpperCAmelCase__ )
294
1
"""simple docstring""" import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _snake_case = logging.get_logger(__name__) _snake_case = {'tokenizer_file': 'tokenizer.json'} _snake_case = { 'tokenizer_file': { 'bigscience/tokenizer': 'https://huggingface.co/bigscience/tokenizer/blob/main/tokenizer.json', 'bigscience/bloom-560m': 'https://huggingface.co/bigscience/bloom-560m/blob/main/tokenizer.json', 'bigscience/bloom-1b1': 'https://huggingface.co/bigscience/bloom-1b1/blob/main/tokenizer.json', 'bigscience/bloom-1b7': 'https://huggingface.co/bigscience/bloom-1b7/blob/main/tokenizer.json', 'bigscience/bloom-3b': 'https://huggingface.co/bigscience/bloom-3b/blob/main/tokenizer.json', 'bigscience/bloom-7b1': 'https://huggingface.co/bigscience/bloom-7b1/blob/main/tokenizer.json', 'bigscience/bloom': 'https://huggingface.co/bigscience/bloom/blob/main/tokenizer.json', }, } class UpperCamelCase ( snake_case_ ): UpperCamelCase : Dict = VOCAB_FILES_NAMES UpperCamelCase : str = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase : List[str] = ['''input_ids''', '''attention_mask'''] UpperCamelCase : str = None def __init__( self : Union[str, Any] , UpperCAmelCase__ : Tuple=None , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : int="<unk>" , UpperCAmelCase__ : Dict="<s>" , UpperCAmelCase__ : List[Any]="</s>" , UpperCAmelCase__ : Any="<pad>" , UpperCAmelCase__ : int=False , UpperCAmelCase__ : Union[str, Any]=False , **UpperCAmelCase__ : Any , ) -> Optional[Any]: super().__init__( UpperCAmelCase__ , UpperCAmelCase__ , tokenizer_file=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , bos_token=UpperCAmelCase__ , eos_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , add_prefix_space=UpperCAmelCase__ , clean_up_tokenization_spaces=UpperCAmelCase__ , **UpperCAmelCase__ , ) _a : List[Any] = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""" , UpperCAmelCase__ ) != add_prefix_space: _a : Optional[Any] = getattr(UpperCAmelCase__ , pre_tok_state.pop("""type""" ) ) _a : Optional[Any] = add_prefix_space _a : Union[str, Any] = pre_tok_class(**UpperCAmelCase__ ) _a : Optional[Any] = add_prefix_space def _lowercase ( self : int , *UpperCAmelCase__ : int , **UpperCAmelCase__ : Tuple ) -> BatchEncoding: _a : List[str] = kwargs.get("""is_split_into_words""" , UpperCAmelCase__ ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( f"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with""" """ pretokenized inputs.""" ) return super()._batch_encode_plus(*UpperCAmelCase__ , **UpperCAmelCase__ ) def _lowercase ( self : List[str] , *UpperCAmelCase__ : str , **UpperCAmelCase__ : List[str] ) -> BatchEncoding: _a : List[Any] = kwargs.get("""is_split_into_words""" , UpperCAmelCase__ ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( f"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with""" """ pretokenized inputs.""" ) return super()._encode_plus(*UpperCAmelCase__ , **UpperCAmelCase__ ) def _lowercase ( self : int , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ) -> Tuple[str]: _a : Optional[int] = self._tokenizer.model.save(UpperCAmelCase__ , name=UpperCAmelCase__ ) return tuple(UpperCAmelCase__ ) def _lowercase ( self : str , UpperCAmelCase__ : "Conversation" ) -> List[int]: _a : List[str] = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) + [self.eos_token_id] ) if len(UpperCAmelCase__ ) > self.model_max_length: _a : Union[str, Any] = input_ids[-self.model_max_length :] return input_ids
294
"""simple docstring""" import unittest import numpy as np def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = None , ): '''simple docstring''' _a : List[Any] = np.shape(UpperCamelCase__ ) _a : Any = np.shape(UpperCamelCase__ ) _a : Union[str, Any] = np.shape(UpperCamelCase__ ) if shape_a[0] != shape_b[0]: _a : int = ( """Expected the same number of rows for A and B. """ F"""Instead found A of size {shape_a} and B of size {shape_b}""" ) raise ValueError(UpperCamelCase__ ) if shape_b[1] != shape_c[1]: _a : Tuple = ( """Expected the same number of columns for B and C. """ F"""Instead found B of size {shape_b} and C of size {shape_c}""" ) raise ValueError(UpperCamelCase__ ) _a : int = pseudo_inv if a_inv is None: try: _a : Optional[int] = np.linalg.inv(UpperCamelCase__ ) except np.linalg.LinAlgError: raise ValueError( """Input matrix A is not invertible. Cannot compute Schur complement.""" ) return mat_c - mat_b.T @ a_inv @ mat_b class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : int ) -> None: _a : str = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) _a : Tuple = np.array([[0, 3], [3, 0], [2, 3]] ) _a : Optional[int] = np.array([[2, 1], [6, 3]] ) _a : Optional[Any] = schur_complement(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) _a : Union[str, Any] = np.block([[a, b], [b.T, c]] ) _a : int = np.linalg.det(UpperCAmelCase__ ) _a : Union[str, Any] = np.linalg.det(UpperCAmelCase__ ) _a : List[Any] = np.linalg.det(UpperCAmelCase__ ) self.assertAlmostEqual(UpperCAmelCase__ , det_a * det_s ) def _lowercase ( self : int ) -> None: _a : Optional[int] = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) _a : Optional[int] = np.array([[0, 3], [3, 0], [2, 3]] ) _a : Union[str, Any] = np.array([[2, 1], [6, 3]] ) with self.assertRaises(UpperCAmelCase__ ): schur_complement(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) def _lowercase ( self : List[Any] ) -> None: _a : Any = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) _a : Dict = np.array([[0, 3], [3, 0], [2, 3]] ) _a : List[Any] = np.array([[2, 1, 3], [6, 3, 5]] ) with self.assertRaises(UpperCAmelCase__ ): schur_complement(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) if __name__ == "__main__": import doctest doctest.testmod() unittest.main()
294
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { 'kssteven/ibert-roberta-base': 'https://huggingface.co/kssteven/ibert-roberta-base/resolve/main/config.json', 'kssteven/ibert-roberta-large': 'https://huggingface.co/kssteven/ibert-roberta-large/resolve/main/config.json', 'kssteven/ibert-roberta-large-mnli': ( 'https://huggingface.co/kssteven/ibert-roberta-large-mnli/resolve/main/config.json' ), } class UpperCamelCase ( snake_case_ ): UpperCamelCase : List[str] = '''ibert''' def __init__( self : Tuple , UpperCAmelCase__ : Dict=30522 , UpperCAmelCase__ : str=768 , UpperCAmelCase__ : Dict=12 , UpperCAmelCase__ : str=12 , UpperCAmelCase__ : Tuple=3072 , UpperCAmelCase__ : Optional[int]="gelu" , UpperCAmelCase__ : Dict=0.1 , UpperCAmelCase__ : int=0.1 , UpperCAmelCase__ : List[Any]=512 , UpperCAmelCase__ : List[str]=2 , UpperCAmelCase__ : Union[str, Any]=0.0_2 , UpperCAmelCase__ : int=1E-12 , UpperCAmelCase__ : Optional[Any]=1 , UpperCAmelCase__ : Dict=0 , UpperCAmelCase__ : Tuple=2 , UpperCAmelCase__ : Optional[int]="absolute" , UpperCAmelCase__ : Union[str, Any]=False , UpperCAmelCase__ : str="none" , **UpperCAmelCase__ : str , ) -> Tuple: super().__init__(pad_token_id=UpperCAmelCase__ , bos_token_id=UpperCAmelCase__ , eos_token_id=UpperCAmelCase__ , **UpperCAmelCase__ ) _a : Optional[Any] = vocab_size _a : List[str] = hidden_size _a : Any = num_hidden_layers _a : Tuple = num_attention_heads _a : int = hidden_act _a : Optional[Any] = intermediate_size _a : Tuple = hidden_dropout_prob _a : Dict = attention_probs_dropout_prob _a : Any = max_position_embeddings _a : List[str] = type_vocab_size _a : List[str] = initializer_range _a : List[Any] = layer_norm_eps _a : Dict = position_embedding_type _a : Optional[int] = quant_mode _a : List[Any] = force_dequant class UpperCamelCase ( snake_case_ ): @property def _lowercase ( self : Optional[int] ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": _a : Any = {0: """batch""", 1: """choice""", 2: """sequence"""} else: _a : List[Any] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
294
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _snake_case = logging.get_logger(__name__) _snake_case = torch.device('cpu') def lowerCAmelCase__ ( ): '''simple docstring''' _a : int = """http://images.cocodataset.org/val2017/000000039769.jpg""" _a : Dict = Image.open(requests.get(UpperCamelCase__ , stream=UpperCamelCase__ ).raw ) return im def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' if swiftformer_name == "swiftformer_xs": return torch.tensor([-2.17_03e00, 2.11_07e00, -2.08_11e00, 8.86_85e-01, 2.43_60e-01] ) elif swiftformer_name == "swiftformer_s": return torch.tensor([3.96_36e-01, 2.34_78e-01, -1.69_63e00, -1.73_81e00, -8.63_37e-01] ) elif swiftformer_name == "swiftformer_l1": return torch.tensor([-4.27_68e-01, -4.74_29e-01, -1.08_97e00, -1.02_48e00, 3.55_23e-02] ) elif swiftformer_name == "swiftformer_l3": return torch.tensor([-2.53_30e-01, 2.42_11e-01, -6.01_85e-01, -8.27_89e-01, -6.04_46e-02] ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Any = dct.pop(UpperCamelCase__ ) _a : Dict = val def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : Tuple = [] for k in state_dict.keys(): _a : Any = k if ".pwconv" in k: _a : int = k_new.replace(""".pwconv""" , """.point_wise_conv""" ) if ".dwconv" in k: _a : List[str] = k_new.replace(""".dwconv""" , """.depth_wise_conv""" ) if ".Proj." in k: _a : Optional[int] = k_new.replace(""".Proj.""" , """.proj.""" ) if "patch_embed" in k_new: _a : Tuple = k_new.replace("""patch_embed""" , """swiftformer.patch_embed.patch_embedding""" ) if "network" in k_new: _a : int = k_new.split(""".""" ) if ls[2].isdigit(): _a : Union[str, Any] = """swiftformer.encoder.network.""" + ls[1] + """.blocks.""" + ls[2] + """.""" + """.""".join(ls[3:] ) else: _a : Tuple = k_new.replace("""network""" , """swiftformer.encoder.network""" ) rename_keys.append((k, k_new) ) return rename_keys @torch.no_grad() def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Tuple = SwiftFormerConfig() # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size _a : Optional[int] = 1_0_0_0 _a : Optional[Any] = """huggingface/label-files""" _a : Optional[Any] = """imagenet-1k-id2label.json""" _a : List[str] = json.load(open(hf_hub_download(UpperCamelCase__ , UpperCamelCase__ , repo_type="""dataset""" ) , """r""" ) ) _a : Optional[Any] = {int(UpperCamelCase__ ): v for k, v in idalabel.items()} _a : Dict = idalabel _a : Optional[int] = {v: k for k, v in idalabel.items()} # size of the architecture if swiftformer_name == "swiftformer_xs": _a : Any = [3, 3, 6, 4] _a : int = [4_8, 5_6, 1_1_2, 2_2_0] elif swiftformer_name == "swiftformer_s": _a : Any = [3, 3, 9, 6] _a : List[str] = [4_8, 6_4, 1_6_8, 2_2_4] elif swiftformer_name == "swiftformer_l1": _a : List[Any] = [4, 3, 1_0, 5] _a : Optional[int] = [4_8, 9_6, 1_9_2, 3_8_4] elif swiftformer_name == "swiftformer_l3": _a : List[Any] = [4, 4, 1_2, 6] _a : Optional[Any] = [6_4, 1_2_8, 3_2_0, 5_1_2] # load state_dict of original model, remove and rename some keys if original_ckpt: if original_ckpt.startswith("""https""" ): _a : Tuple = torch.hub.load_state_dict_from_url(UpperCamelCase__ , map_location="""cpu""" , check_hash=UpperCamelCase__ ) else: _a : Dict = torch.load(UpperCamelCase__ , map_location="""cpu""" ) _a : int = checkpoint _a : Optional[Any] = create_rename_keys(UpperCamelCase__ ) for rename_key_src, rename_key_dest in rename_keys: rename_key(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # load HuggingFace model _a : Any = SwiftFormerForImageClassification(UpperCamelCase__ ).eval() hf_model.load_state_dict(UpperCamelCase__ ) # prepare test inputs _a : Any = prepare_img() _a : Union[str, Any] = ViTImageProcessor.from_pretrained("""preprocessor_config""" ) _a : Optional[int] = processor(images=UpperCamelCase__ , return_tensors="""pt""" ) # compare outputs from both models _a : Dict = get_expected_output(UpperCamelCase__ ) _a : int = hf_model(inputs["""pixel_values"""] ).logits assert hf_logits.shape == torch.Size([1, 1_0_0_0] ) assert torch.allclose(hf_logits[0, 0:5] , UpperCamelCase__ , atol=1e-3 ) Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) print(F"""Saving model {swiftformer_name} to {pytorch_dump_folder_path}""" ) hf_model.save_pretrained(UpperCamelCase__ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '--swiftformer_name', default='swiftformer_xs', choices=['swiftformer_xs', 'swiftformer_s', 'swiftformer_l1', 'swiftformer_l3'], type=str, help='Name of the SwiftFormer model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default='./converted_outputs/', type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument('--original_ckpt', default=None, type=str, help='Path to the original model checkpoint.') _snake_case = parser.parse_args() convert_swiftformer_checkpoint(args.swiftformer_name, args.pytorch_dump_folder_path, args.original_ckpt)
294
1
"""simple docstring""" from dataclasses import dataclass from typing import Optional import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .modeling_utils import ModelMixin @dataclass class UpperCamelCase ( snake_case_ ): UpperCamelCase : torch.FloatTensor class UpperCamelCase ( snake_case_ , snake_case_ ): @register_to_config def __init__( self : int , UpperCAmelCase__ : int = 16 , UpperCAmelCase__ : int = 88 , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : int = 1 , UpperCAmelCase__ : float = 0.0 , UpperCAmelCase__ : int = 32 , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : bool = False , UpperCAmelCase__ : Optional[int] = None , UpperCAmelCase__ : str = "geglu" , UpperCAmelCase__ : bool = True , UpperCAmelCase__ : bool = True , ) -> str: super().__init__() _a : List[str] = num_attention_heads _a : str = attention_head_dim _a : List[Any] = num_attention_heads * attention_head_dim _a : int = in_channels _a : str = torch.nn.GroupNorm(num_groups=UpperCAmelCase__ , num_channels=UpperCAmelCase__ , eps=1E-6 , affine=UpperCAmelCase__ ) _a : Tuple = nn.Linear(UpperCAmelCase__ , UpperCAmelCase__ ) # 3. Define transformers blocks _a : Dict = nn.ModuleList( [ BasicTransformerBlock( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , dropout=UpperCAmelCase__ , cross_attention_dim=UpperCAmelCase__ , activation_fn=UpperCAmelCase__ , attention_bias=UpperCAmelCase__ , double_self_attention=UpperCAmelCase__ , norm_elementwise_affine=UpperCAmelCase__ , ) for d in range(UpperCAmelCase__ ) ] ) _a : Any = nn.Linear(UpperCAmelCase__ , UpperCAmelCase__ ) def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : List[Any]=None , UpperCAmelCase__ : Optional[Any]=None , UpperCAmelCase__ : int=None , UpperCAmelCase__ : Union[str, Any]=1 , UpperCAmelCase__ : int=None , UpperCAmelCase__ : bool = True , ) -> Dict: _a , _a , _a , _a : List[Any] = hidden_states.shape _a : str = batch_frames // num_frames _a : Any = hidden_states _a : Dict = hidden_states[None, :].reshape(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) _a : Tuple = hidden_states.permute(0 , 2 , 1 , 3 , 4 ) _a : Tuple = self.norm(UpperCAmelCase__ ) _a : Optional[int] = hidden_states.permute(0 , 3 , 4 , 2 , 1 ).reshape(batch_size * height * width , UpperCAmelCase__ , UpperCAmelCase__ ) _a : Union[str, Any] = self.proj_in(UpperCAmelCase__ ) # 2. Blocks for block in self.transformer_blocks: _a : Dict = block( UpperCAmelCase__ , encoder_hidden_states=UpperCAmelCase__ , timestep=UpperCAmelCase__ , cross_attention_kwargs=UpperCAmelCase__ , class_labels=UpperCAmelCase__ , ) # 3. Output _a : Tuple = self.proj_out(UpperCAmelCase__ ) _a : Optional[int] = ( hidden_states[None, None, :] .reshape(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) .permute(0 , 3 , 4 , 1 , 2 ) .contiguous() ) _a : List[str] = hidden_states.reshape(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) _a : Optional[int] = hidden_states + residual if not return_dict: return (output,) return TransformerTemporalModelOutput(sample=UpperCAmelCase__ )
294
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case = { 'configuration_perceiver': ['PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PerceiverConfig', 'PerceiverOnnxConfig'], 'tokenization_perceiver': ['PerceiverTokenizer'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ['PerceiverFeatureExtractor'] _snake_case = ['PerceiverImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST', 'PerceiverForImageClassificationConvProcessing', 'PerceiverForImageClassificationFourier', 'PerceiverForImageClassificationLearned', 'PerceiverForMaskedLM', 'PerceiverForMultimodalAutoencoding', 'PerceiverForOpticalFlow', 'PerceiverForSequenceClassification', 'PerceiverLayer', 'PerceiverModel', 'PerceiverPreTrainedModel', ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
294
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _snake_case = logging.get_logger(__name__) _snake_case = { 'shi-labs/nat-mini-in1k-224': 'https://huggingface.co/shi-labs/nat-mini-in1k-224/resolve/main/config.json', # See all Nat models at https://huggingface.co/models?filter=nat } class UpperCamelCase ( snake_case_ , snake_case_ ): UpperCamelCase : Any = '''nat''' UpperCamelCase : List[Any] = { '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers''', } def __init__( self : Union[str, Any] , UpperCAmelCase__ : List[Any]=4 , UpperCAmelCase__ : int=3 , UpperCAmelCase__ : Union[str, Any]=64 , UpperCAmelCase__ : str=[3, 4, 6, 5] , UpperCAmelCase__ : Optional[Any]=[2, 4, 8, 16] , UpperCAmelCase__ : str=7 , UpperCAmelCase__ : Optional[Any]=3.0 , UpperCAmelCase__ : int=True , UpperCAmelCase__ : str=0.0 , UpperCAmelCase__ : str=0.0 , UpperCAmelCase__ : str=0.1 , UpperCAmelCase__ : List[Any]="gelu" , UpperCAmelCase__ : List[str]=0.0_2 , UpperCAmelCase__ : str=1E-5 , UpperCAmelCase__ : List[Any]=0.0 , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : Union[str, Any]=None , **UpperCAmelCase__ : Union[str, Any] , ) -> List[str]: super().__init__(**UpperCAmelCase__ ) _a : Any = patch_size _a : Union[str, Any] = num_channels _a : List[str] = embed_dim _a : List[Any] = depths _a : int = len(UpperCAmelCase__ ) _a : Optional[Any] = num_heads _a : int = kernel_size _a : List[str] = mlp_ratio _a : Dict = qkv_bias _a : List[Any] = hidden_dropout_prob _a : Optional[Any] = attention_probs_dropout_prob _a : Optional[Any] = drop_path_rate _a : Dict = hidden_act _a : Dict = layer_norm_eps _a : List[str] = initializer_range # we set the hidden_size attribute in order to make Nat work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model _a : List[Any] = int(embed_dim * 2 ** (len(UpperCAmelCase__ ) - 1) ) _a : List[Any] = layer_scale_init_value _a : Dict = ["""stem"""] + [f"""stage{idx}""" for idx in range(1 , len(UpperCAmelCase__ ) + 1 )] _a , _a : List[Any] = get_aligned_output_features_output_indices( out_features=UpperCAmelCase__ , out_indices=UpperCAmelCase__ , stage_names=self.stage_names )
294
"""simple docstring""" import cva import numpy as np class UpperCamelCase : def __init__( self : Optional[int] , UpperCAmelCase__ : float , UpperCAmelCase__ : int ) -> Dict: if k in (0.0_4, 0.0_6): _a : List[str] = k _a : List[Any] = window_size else: raise ValueError("""invalid k value""" ) def __str__( self : Dict ) -> str: return str(self.k ) def _lowercase ( self : int , UpperCAmelCase__ : str ) -> tuple[cva.Mat, list[list[int]]]: _a : Dict = cva.imread(UpperCAmelCase__ , 0 ) _a , _a : List[Any] = img.shape _a : list[list[int]] = [] _a : List[Any] = img.copy() _a : int = cva.cvtColor(UpperCAmelCase__ , cva.COLOR_GRAY2RGB ) _a , _a : Any = np.gradient(UpperCAmelCase__ ) _a : Tuple = dx**2 _a : Union[str, Any] = dy**2 _a : Union[str, Any] = dx * dy _a : int = 0.0_4 _a : List[str] = self.window_size // 2 for y in range(UpperCAmelCase__ , h - offset ): for x in range(UpperCAmelCase__ , w - offset ): _a : str = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _a : List[Any] = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _a : Tuple = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _a : Any = (wxx * wyy) - (wxy**2) _a : Tuple = wxx + wyy _a : Any = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 255 ) return color_img, corner_list if __name__ == "__main__": _snake_case = HarrisCorner(0.04, 3) _snake_case , _snake_case = edge_detect.detect('path_to_image') cva.imwrite('detect.png', color_img)
294
1
"""simple docstring""" import os import unittest from transformers import BatchEncoding from transformers.models.bert.tokenization_bert import ( BasicTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.models.prophetnet.tokenization_prophetnet import VOCAB_FILES_NAMES, ProphetNetTokenizer from transformers.testing_utils import require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin class UpperCamelCase ( snake_case_ , unittest.TestCase ): UpperCamelCase : Tuple = ProphetNetTokenizer UpperCamelCase : Optional[Any] = False def _lowercase ( self : Optional[int] ) -> str: super().setUp() _a : str = [ """[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest""", ] _a : int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def _lowercase ( self : Union[str, Any] , UpperCAmelCase__ : Any ) -> Optional[Any]: _a : str = """UNwant\u00E9d,running""" _a : Union[str, Any] = """unwanted, running""" return input_text, output_text def _lowercase ( self : List[Any] ) -> Any: _a : int = self.tokenizer_class(self.vocab_file ) _a : List[str] = tokenizer.tokenize("""UNwant\u00E9d,running""" ) self.assertListEqual(UpperCAmelCase__ , ["""un""", """##want""", """##ed""", """,""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase__ ) , [9, 6, 7, 12, 10, 11] ) def _lowercase ( self : List[Any] ) -> List[Any]: _a : List[str] = BasicTokenizer() self.assertListEqual(tokenizer.tokenize("""ah\u535A\u63A8zz""" ) , ["""ah""", """\u535A""", """\u63A8""", """zz"""] ) def _lowercase ( self : Optional[int] ) -> int: _a : Union[str, Any] = BasicTokenizer(do_lower_case=UpperCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""hello""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def _lowercase ( self : Tuple ) -> List[str]: _a : List[Any] = BasicTokenizer(do_lower_case=UpperCAmelCase__ , strip_accents=UpperCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hällo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""h\u00E9llo"""] ) def _lowercase ( self : Optional[int] ) -> Dict: _a : Union[str, Any] = BasicTokenizer(do_lower_case=UpperCAmelCase__ , strip_accents=UpperCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def _lowercase ( self : Dict ) -> Tuple: _a : Optional[int] = BasicTokenizer(do_lower_case=UpperCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def _lowercase ( self : Optional[int] ) -> Union[str, Any]: _a : List[str] = BasicTokenizer(do_lower_case=UpperCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _lowercase ( self : Dict ) -> Optional[Any]: _a : Tuple = BasicTokenizer(do_lower_case=UpperCAmelCase__ , strip_accents=UpperCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HäLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _lowercase ( self : List[str] ) -> Union[str, Any]: _a : int = BasicTokenizer(do_lower_case=UpperCAmelCase__ , strip_accents=UpperCAmelCase__ ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HaLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _lowercase ( self : Dict ) -> Any: _a : Tuple = BasicTokenizer(do_lower_case=UpperCAmelCase__ , never_split=["""[UNK]"""] ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? [UNK]""" ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?""", """[UNK]"""] ) def _lowercase ( self : Union[str, Any] ) -> Union[str, Any]: _a : List[Any] = ["""[UNK]""", """[CLS]""", """[SEP]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing"""] _a : int = {} for i, token in enumerate(UpperCAmelCase__ ): _a : Optional[Any] = i _a : Optional[int] = WordpieceTokenizer(vocab=UpperCAmelCase__ , unk_token="""[UNK]""" ) self.assertListEqual(tokenizer.tokenize("""""" ) , [] ) self.assertListEqual(tokenizer.tokenize("""unwanted running""" ) , ["""un""", """##want""", """##ed""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.tokenize("""unwantedX running""" ) , ["""[UNK]""", """runn""", """##ing"""] ) @require_torch def _lowercase ( self : List[str] ) -> Optional[int]: _a : List[Any] = self.tokenizer_class.from_pretrained("""microsoft/prophetnet-large-uncased""" ) _a : Any = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] _a : Optional[Any] = [1037, 2146, 20423, 2005, 7680, 7849, 3989, 1012, 102] _a : str = tokenizer(UpperCAmelCase__ , padding=UpperCAmelCase__ , return_tensors="""pt""" ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Optional[Any] = list(batch.input_ids.numpy()[0] ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) self.assertEqual((2, 9) , batch.input_ids.shape ) self.assertEqual((2, 9) , batch.attention_mask.shape ) def _lowercase ( self : Optional[int] ) -> Dict: self.assertTrue(_is_whitespace(""" """ ) ) self.assertTrue(_is_whitespace("""\t""" ) ) self.assertTrue(_is_whitespace("""\r""" ) ) self.assertTrue(_is_whitespace("""\n""" ) ) self.assertTrue(_is_whitespace("""\u00A0""" ) ) self.assertFalse(_is_whitespace("""A""" ) ) self.assertFalse(_is_whitespace("""-""" ) ) def _lowercase ( self : Tuple ) -> str: self.assertTrue(_is_control("""\u0005""" ) ) self.assertFalse(_is_control("""A""" ) ) self.assertFalse(_is_control(""" """ ) ) self.assertFalse(_is_control("""\t""" ) ) self.assertFalse(_is_control("""\r""" ) ) def _lowercase ( self : Dict ) -> Optional[int]: self.assertTrue(_is_punctuation("""-""" ) ) self.assertTrue(_is_punctuation("""$""" ) ) self.assertTrue(_is_punctuation("""`""" ) ) self.assertTrue(_is_punctuation(""".""" ) ) self.assertFalse(_is_punctuation("""A""" ) ) self.assertFalse(_is_punctuation(""" """ ) ) @slow def _lowercase ( self : List[Any] ) -> str: _a : Union[str, Any] = self.tokenizer_class.from_pretrained("""microsoft/prophetnet-large-uncased""" ) _a : Optional[Any] = tokenizer.encode("""sequence builders""" , add_special_tokens=UpperCAmelCase__ ) _a : Optional[int] = tokenizer.encode("""multi-sequence build""" , add_special_tokens=UpperCAmelCase__ ) _a : List[str] = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase__ ) _a : int = tokenizer.build_inputs_with_special_tokens(UpperCAmelCase__ , UpperCAmelCase__ ) assert encoded_sentence == text + [102] assert encoded_pair == text + [102] + text_a + [102]
294
"""simple docstring""" def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' # Check if the input is valid if not len(UpperCamelCase__ ) == len(UpperCamelCase__ ) == 3: raise ValueError("""Please enter a valid equation.""" ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError("""Both a & b of two equations can't be zero.""" ) # Extract the coefficients _a , _a , _a : Any = equationa _a , _a , _a : Tuple = equationa # Calculate the determinants of the matrices _a : int = aa * ba - aa * ba _a : str = ca * ba - ca * ba _a : str = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError("""Infinite solutions. (Consistent system)""" ) else: raise ValueError("""No solution. (Inconsistent system)""" ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: _a : Dict = determinant_x / determinant _a : str = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
294
1
"""simple docstring""" import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class UpperCamelCase ( snake_case_ ): UpperCamelCase : str = (DEISMultistepScheduler,) UpperCamelCase : Dict = (('''num_inference_steps''', 25),) def _lowercase ( self : Dict , **UpperCAmelCase__ : Tuple ) -> str: _a : Optional[Any] = { """num_train_timesteps""": 1000, """beta_start""": 0.0_0_0_1, """beta_end""": 0.0_2, """beta_schedule""": """linear""", """solver_order""": 2, } config.update(**UpperCAmelCase__ ) return config def _lowercase ( self : List[Any] , UpperCAmelCase__ : Optional[Any]=0 , **UpperCAmelCase__ : Tuple ) -> Tuple: _a : List[Any] = dict(self.forward_default_kwargs ) _a : Union[str, Any] = kwargs.pop("""num_inference_steps""" , UpperCAmelCase__ ) _a : Union[str, Any] = self.dummy_sample _a : List[str] = 0.1 * sample _a : Optional[int] = [residual + 0.2, residual + 0.1_5, residual + 0.1_0] for scheduler_class in self.scheduler_classes: _a : Optional[Any] = self.get_scheduler_config(**UpperCAmelCase__ ) _a : Tuple = scheduler_class(**UpperCAmelCase__ ) scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residuals _a : int = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(UpperCAmelCase__ ) _a : Union[str, Any] = scheduler_class.from_pretrained(UpperCAmelCase__ ) new_scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residuals _a : Tuple = dummy_past_residuals[: new_scheduler.config.solver_order] _a , _a : List[Any] = sample, sample for t in range(UpperCAmelCase__ , time_step + scheduler.config.solver_order + 1 ): _a : str = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : int = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def _lowercase ( self : List[str] ) -> Tuple: pass def _lowercase ( self : Optional[int] , UpperCAmelCase__ : Tuple=0 , **UpperCAmelCase__ : List[Any] ) -> Tuple: _a : str = dict(self.forward_default_kwargs ) _a : Optional[int] = kwargs.pop("""num_inference_steps""" , UpperCAmelCase__ ) _a : Optional[int] = self.dummy_sample _a : Dict = 0.1 * sample _a : int = [residual + 0.2, residual + 0.1_5, residual + 0.1_0] for scheduler_class in self.scheduler_classes: _a : Any = self.get_scheduler_config() _a : Optional[Any] = scheduler_class(**UpperCAmelCase__ ) scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residuals (must be after setting timesteps) _a : Union[str, Any] = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(UpperCAmelCase__ ) _a : Tuple = scheduler_class.from_pretrained(UpperCAmelCase__ ) # copy over dummy past residuals new_scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residual (must be after setting timesteps) _a : List[str] = dummy_past_residuals[: new_scheduler.config.solver_order] _a : List[Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : Tuple = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def _lowercase ( self : Dict , UpperCAmelCase__ : List[str]=None , **UpperCAmelCase__ : str ) -> Optional[int]: if scheduler is None: _a : Dict = self.scheduler_classes[0] _a : Union[str, Any] = self.get_scheduler_config(**UpperCAmelCase__ ) _a : Dict = scheduler_class(**UpperCAmelCase__ ) _a : Union[str, Any] = self.scheduler_classes[0] _a : List[str] = self.get_scheduler_config(**UpperCAmelCase__ ) _a : int = scheduler_class(**UpperCAmelCase__ ) _a : str = 10 _a : Optional[Any] = self.dummy_model() _a : List[Any] = self.dummy_sample_deter scheduler.set_timesteps(UpperCAmelCase__ ) for i, t in enumerate(scheduler.timesteps ): _a : List[str] = model(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[str] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ).prev_sample return sample def _lowercase ( self : List[str] ) -> Any: _a : str = dict(self.forward_default_kwargs ) _a : Tuple = kwargs.pop("""num_inference_steps""" , UpperCAmelCase__ ) for scheduler_class in self.scheduler_classes: _a : Optional[int] = self.get_scheduler_config() _a : Union[str, Any] = scheduler_class(**UpperCAmelCase__ ) _a : int = self.dummy_sample _a : List[str] = 0.1 * sample if num_inference_steps is not None and hasattr(UpperCAmelCase__ , """set_timesteps""" ): scheduler.set_timesteps(UpperCAmelCase__ ) elif num_inference_steps is not None and not hasattr(UpperCAmelCase__ , """set_timesteps""" ): _a : int = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) _a : Optional[int] = [residual + 0.2, residual + 0.1_5, residual + 0.1_0] _a : Tuple = dummy_past_residuals[: scheduler.config.solver_order] _a : Dict = scheduler.timesteps[5] _a : str = scheduler.timesteps[6] _a : Optional[Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : int = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def _lowercase ( self : int ) -> int: # make sure that iterating over schedulers with same config names gives same results # for defaults _a : List[Any] = DEISMultistepScheduler(**self.get_scheduler_config() ) _a : Dict = self.full_loop(scheduler=UpperCAmelCase__ ) _a : Tuple = torch.mean(torch.abs(UpperCAmelCase__ ) ) assert abs(result_mean.item() - 0.2_3_9_1_6 ) < 1E-3 _a : Optional[int] = DPMSolverSinglestepScheduler.from_config(scheduler.config ) _a : Dict = DPMSolverMultistepScheduler.from_config(scheduler.config ) _a : Optional[Any] = UniPCMultistepScheduler.from_config(scheduler.config ) _a : List[str] = DEISMultistepScheduler.from_config(scheduler.config ) _a : List[Any] = self.full_loop(scheduler=UpperCAmelCase__ ) _a : Tuple = torch.mean(torch.abs(UpperCAmelCase__ ) ) assert abs(result_mean.item() - 0.2_3_9_1_6 ) < 1E-3 def _lowercase ( self : Dict ) -> int: for timesteps in [25, 50, 100, 999, 1000]: self.check_over_configs(num_train_timesteps=UpperCAmelCase__ ) def _lowercase ( self : Union[str, Any] ) -> Union[str, Any]: self.check_over_configs(thresholding=UpperCAmelCase__ ) for order in [1, 2, 3]: for solver_type in ["logrho"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=UpperCAmelCase__ , prediction_type=UpperCAmelCase__ , sample_max_value=UpperCAmelCase__ , algorithm_type="""deis""" , solver_order=UpperCAmelCase__ , solver_type=UpperCAmelCase__ , ) def _lowercase ( self : Dict ) -> Optional[Any]: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=UpperCAmelCase__ ) def _lowercase ( self : List[Any] ) -> Optional[int]: for algorithm_type in ["deis"]: for solver_type in ["logrho"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=UpperCAmelCase__ , solver_type=UpperCAmelCase__ , prediction_type=UpperCAmelCase__ , algorithm_type=UpperCAmelCase__ , ) _a : int = self.full_loop( solver_order=UpperCAmelCase__ , solver_type=UpperCAmelCase__ , prediction_type=UpperCAmelCase__ , algorithm_type=UpperCAmelCase__ , ) assert not torch.isnan(UpperCAmelCase__ ).any(), "Samples have nan numbers" def _lowercase ( self : List[Any] ) -> List[Any]: self.check_over_configs(lower_order_final=UpperCAmelCase__ ) self.check_over_configs(lower_order_final=UpperCAmelCase__ ) def _lowercase ( self : Optional[int] ) -> Union[str, Any]: for num_inference_steps in [1, 2, 3, 5, 10, 50, 100, 999, 1000]: self.check_over_forward(num_inference_steps=UpperCAmelCase__ , time_step=0 ) def _lowercase ( self : int ) -> int: _a : Union[str, Any] = self.full_loop() _a : Optional[Any] = torch.mean(torch.abs(UpperCAmelCase__ ) ) assert abs(result_mean.item() - 0.2_3_9_1_6 ) < 1E-3 def _lowercase ( self : int ) -> Tuple: _a : List[Any] = self.full_loop(prediction_type="""v_prediction""" ) _a : Any = torch.mean(torch.abs(UpperCAmelCase__ ) ) assert abs(result_mean.item() - 0.0_9_1 ) < 1E-3 def _lowercase ( self : Optional[Any] ) -> Optional[int]: _a : Tuple = self.scheduler_classes[0] _a : Any = self.get_scheduler_config(thresholding=UpperCAmelCase__ , dynamic_thresholding_ratio=0 ) _a : Tuple = scheduler_class(**UpperCAmelCase__ ) _a : int = 10 _a : Dict = self.dummy_model() _a : List[Any] = self.dummy_sample_deter.half() scheduler.set_timesteps(UpperCAmelCase__ ) for i, t in enumerate(scheduler.timesteps ): _a : List[str] = model(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ).prev_sample assert sample.dtype == torch.floataa
294
"""simple docstring""" _snake_case = '\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n' _snake_case = [{'type': 'code', 'content': INSTALL_CONTENT}] _snake_case = { '{processor_class}': 'FakeProcessorClass', '{model_class}': 'FakeModelClass', '{object_class}': 'FakeObjectClass', }
294
1
"""simple docstring""" def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' global f # a global dp table for knapsack if f[i][j] < 0: if j < wt[i - 1]: _a : Optional[Any] = mf_knapsack(i - 1 , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else: _a : Any = max( mf_knapsack(i - 1 , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) , mf_knapsack(i - 1 , UpperCamelCase__ , UpperCamelCase__ , j - wt[i - 1] ) + val[i - 1] , ) _a : int = val return f[i][j] def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Optional[int] = [[0] * (w + 1) for _ in range(n + 1 )] for i in range(1 , n + 1 ): for w_ in range(1 , w + 1 ): if wt[i - 1] <= w_: _a : int = max(val[i - 1] + dp[i - 1][w_ - wt[i - 1]] , dp[i - 1][w_] ) else: _a : Any = dp[i - 1][w_] return dp[n][w_], dp def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' if not (isinstance(UpperCamelCase__ , (list, tuple) ) and isinstance(UpperCamelCase__ , (list, tuple) )): raise ValueError( """Both the weights and values vectors must be either lists or tuples""" ) _a : Union[str, Any] = len(UpperCamelCase__ ) if num_items != len(UpperCamelCase__ ): _a : Union[str, Any] = ( """The number of weights must be the same as the number of values.\n""" F"""But got {num_items} weights and {len(UpperCamelCase__ )} values""" ) raise ValueError(UpperCamelCase__ ) for i in range(UpperCamelCase__ ): if not isinstance(wt[i] , UpperCamelCase__ ): _a : Tuple = ( """All weights must be integers but got weight of """ F"""type {type(wt[i] )} at index {i}""" ) raise TypeError(UpperCamelCase__ ) _a , _a : List[Any] = knapsack(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) _a : set = set() _construct_solution(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) return optimal_val, example_optional_set def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' # for the current item i at a maximum weight j to be part of an optimal subset, # the optimal value at (i, j) must be greater than the optimal value at (i-1, j). # where i - 1 means considering only the previous items at the given maximum weight if i > 0 and j > 0: if dp[i - 1][j] == dp[i][j]: _construct_solution(UpperCamelCase__ , UpperCamelCase__ , i - 1 , UpperCamelCase__ , UpperCamelCase__ ) else: optimal_set.add(UpperCamelCase__ ) _construct_solution(UpperCamelCase__ , UpperCamelCase__ , i - 1 , j - wt[i - 1] , UpperCamelCase__ ) if __name__ == "__main__": _snake_case = [3, 2, 4, 4] _snake_case = [4, 3, 2, 3] _snake_case = 4 _snake_case = 6 _snake_case = [[0] * (w + 1)] + [[0] + [-1] * (w + 1) for _ in range(n + 1)] _snake_case , _snake_case = knapsack(w, wt, val, n) print(optimal_solution) print(mf_knapsack(n, wt, val, w)) # switched the n and w # testing the dynamic programming problem with example # the optimal subset for the above example are items 3 and 4 _snake_case , _snake_case = knapsack_with_example_solution(w, wt, val) assert optimal_solution == 8 assert optimal_subset == {3, 4} print('optimal_value = ', optimal_solution) print('An optimal subset corresponding to the optimal value', optimal_subset)
294
"""simple docstring""" import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json', } class UpperCamelCase ( snake_case_ ): UpperCamelCase : int = '''mvp''' UpperCamelCase : Union[str, Any] = ['''past_key_values'''] UpperCamelCase : Any = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : List[str] , UpperCAmelCase__ : List[str]=50267 , UpperCAmelCase__ : Optional[Any]=1024 , UpperCAmelCase__ : Tuple=12 , UpperCAmelCase__ : Optional[Any]=4096 , UpperCAmelCase__ : int=16 , UpperCAmelCase__ : Tuple=12 , UpperCAmelCase__ : int=4096 , UpperCAmelCase__ : List[Any]=16 , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : Tuple="gelu" , UpperCAmelCase__ : Union[str, Any]=1024 , UpperCAmelCase__ : List[str]=0.1 , UpperCAmelCase__ : Any=0.0 , UpperCAmelCase__ : Dict=0.0 , UpperCAmelCase__ : Tuple=0.0_2 , UpperCAmelCase__ : Tuple=0.0 , UpperCAmelCase__ : Optional[Any]=False , UpperCAmelCase__ : int=True , UpperCAmelCase__ : Tuple=1 , UpperCAmelCase__ : Dict=0 , UpperCAmelCase__ : Union[str, Any]=2 , UpperCAmelCase__ : Optional[int]=True , UpperCAmelCase__ : Tuple=2 , UpperCAmelCase__ : Any=2 , UpperCAmelCase__ : Optional[Any]=False , UpperCAmelCase__ : Dict=100 , UpperCAmelCase__ : Union[str, Any]=800 , **UpperCAmelCase__ : Dict , ) -> List[Any]: _a : Any = vocab_size _a : Any = max_position_embeddings _a : Union[str, Any] = d_model _a : List[str] = encoder_ffn_dim _a : List[Any] = encoder_layers _a : Dict = encoder_attention_heads _a : Tuple = decoder_ffn_dim _a : List[Any] = decoder_layers _a : Optional[Any] = decoder_attention_heads _a : Optional[Any] = dropout _a : str = attention_dropout _a : Dict = activation_dropout _a : Any = activation_function _a : Tuple = init_std _a : Dict = encoder_layerdrop _a : Optional[int] = decoder_layerdrop _a : Optional[Any] = classifier_dropout _a : List[Any] = use_cache _a : Dict = encoder_layers _a : str = scale_embedding # scale factor will be sqrt(d_model) if True _a : int = use_prompt _a : Dict = prompt_length _a : Dict = prompt_mid_dim super().__init__( pad_token_id=UpperCAmelCase__ , bos_token_id=UpperCAmelCase__ , eos_token_id=UpperCAmelCase__ , is_encoder_decoder=UpperCAmelCase__ , decoder_start_token_id=UpperCAmelCase__ , forced_eos_token_id=UpperCAmelCase__ , **UpperCAmelCase__ , ) if self.forced_bos_token_id is None and kwargs.get("""force_bos_token_to_be_generated""" , UpperCAmelCase__ ): _a : List[str] = self.bos_token_id warnings.warn( f"""Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. """ """The config can simply be saved and uploaded again to be fixed.""" )
294
1
"""simple docstring""" _snake_case = {'a': ['c', 'b'], 'b': ['d', 'e'], 'c': [], 'd': [], 'e': []} _snake_case = ['a', 'b', 'c', 'd', 'e'] def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : int = start # add current to visited visited.append(UpperCamelCase__ ) _a : Any = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: _a : Union[str, Any] = topological_sort(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # if all neighbors visited add current to sort sort.append(UpperCamelCase__ ) # if all vertices haven't been visited select a new one to visit if len(UpperCamelCase__ ) != len(UpperCamelCase__ ): for vertice in vertices: if vertice not in visited: _a : Optional[Any] = topological_sort(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # return sort return sort if __name__ == "__main__": _snake_case = topological_sort('a', [], []) print(sort)
294
"""simple docstring""" import logging import os from typing import List, TextIO, Union from conllu import parse_incr from utils_ner import InputExample, Split, TokenClassificationTask _snake_case = logging.getLogger(__name__) class UpperCamelCase ( snake_case_ ): def __init__( self : Optional[Any] , UpperCAmelCase__ : Optional[int]=-1 ) -> Tuple: # in NER datasets, the last column is usually reserved for NER label _a : Optional[int] = label_idx def _lowercase ( self : Any , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Union[Split, str] ) -> List[InputExample]: if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): _a : Any = mode.value _a : Optional[int] = os.path.join(UpperCAmelCase__ , f"""{mode}.txt""" ) _a : int = 1 _a : int = [] with open(UpperCAmelCase__ , encoding="""utf-8""" ) as f: _a : str = [] _a : str = [] for line in f: if line.startswith("""-DOCSTART-""" ) or line == "" or line == "\n": if words: examples.append(InputExample(guid=f"""{mode}-{guid_index}""" , words=UpperCAmelCase__ , labels=UpperCAmelCase__ ) ) guid_index += 1 _a : List[str] = [] _a : str = [] else: _a : List[Any] = line.split(""" """ ) words.append(splits[0] ) if len(UpperCAmelCase__ ) > 1: labels.append(splits[self.label_idx].replace("""\n""" , """""" ) ) else: # Examples could have no label for mode = "test" labels.append("""O""" ) if words: examples.append(InputExample(guid=f"""{mode}-{guid_index}""" , words=UpperCAmelCase__ , labels=UpperCAmelCase__ ) ) return examples def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : List ) -> Union[str, Any]: _a : List[str] = 0 for line in test_input_reader: if line.startswith("""-DOCSTART-""" ) or line == "" or line == "\n": writer.write(UpperCAmelCase__ ) if not preds_list[example_id]: example_id += 1 elif preds_list[example_id]: _a : int = line.split()[0] + """ """ + preds_list[example_id].pop(0 ) + """\n""" writer.write(UpperCAmelCase__ ) else: logger.warning("""Maximum sequence length exceeded: No prediction for '%s'.""" , line.split()[0] ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : str ) -> List[str]: if path: with open(UpperCAmelCase__ , """r""" ) as f: _a : List[Any] = f.read().splitlines() if "O" not in labels: _a : Union[str, Any] = ["""O"""] + labels return labels else: return ["O", "B-MISC", "I-MISC", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC"] class UpperCamelCase ( snake_case_ ): def __init__( self : Union[str, Any] ) -> List[str]: # in CONLL2003 dataset chunk column is second-to-last super().__init__(label_idx=-2 ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : str ) -> List[str]: if path: with open(UpperCAmelCase__ , """r""" ) as f: _a : Optional[int] = f.read().splitlines() if "O" not in labels: _a : Optional[Any] = ["""O"""] + labels return labels else: return [ "O", "B-ADVP", "B-INTJ", "B-LST", "B-PRT", "B-NP", "B-SBAR", "B-VP", "B-ADJP", "B-CONJP", "B-PP", "I-ADVP", "I-INTJ", "I-LST", "I-PRT", "I-NP", "I-SBAR", "I-VP", "I-ADJP", "I-CONJP", "I-PP", ] class UpperCamelCase ( snake_case_ ): def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Union[Split, str] ) -> List[InputExample]: if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): _a : List[Any] = mode.value _a : Union[str, Any] = os.path.join(UpperCAmelCase__ , f"""{mode}.txt""" ) _a : List[str] = 1 _a : Optional[Any] = [] with open(UpperCAmelCase__ , encoding="""utf-8""" ) as f: for sentence in parse_incr(UpperCAmelCase__ ): _a : List[Any] = [] _a : Any = [] for token in sentence: words.append(token["""form"""] ) labels.append(token["""upos"""] ) assert len(UpperCAmelCase__ ) == len(UpperCAmelCase__ ) if words: examples.append(InputExample(guid=f"""{mode}-{guid_index}""" , words=UpperCAmelCase__ , labels=UpperCAmelCase__ ) ) guid_index += 1 return examples def _lowercase ( self : Tuple , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : TextIO , UpperCAmelCase__ : List ) -> Dict: _a : Optional[Any] = 0 for sentence in parse_incr(UpperCAmelCase__ ): _a : List[str] = preds_list[example_id] _a : str = """""" for token in sentence: out += f"""{token['form']} ({token['upos']}|{s_p.pop(0 )}) """ out += "\n" writer.write(UpperCAmelCase__ ) example_id += 1 def _lowercase ( self : List[str] , UpperCAmelCase__ : str ) -> List[str]: if path: with open(UpperCAmelCase__ , """r""" ) as f: return f.read().splitlines() else: return [ "ADJ", "ADP", "ADV", "AUX", "CCONJ", "DET", "INTJ", "NOUN", "NUM", "PART", "PRON", "PROPN", "PUNCT", "SCONJ", "SYM", "VERB", "X", ]
294
1
"""simple docstring""" def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' # Check if the input is valid if not len(UpperCamelCase__ ) == len(UpperCamelCase__ ) == 3: raise ValueError("""Please enter a valid equation.""" ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError("""Both a & b of two equations can't be zero.""" ) # Extract the coefficients _a , _a , _a : Any = equationa _a , _a , _a : Tuple = equationa # Calculate the determinants of the matrices _a : int = aa * ba - aa * ba _a : str = ca * ba - ca * ba _a : str = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError("""Infinite solutions. (Consistent system)""" ) else: raise ValueError("""No solution. (Inconsistent system)""" ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: _a : Dict = determinant_x / determinant _a : str = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
294
"""simple docstring""" from __future__ import annotations import time import numpy as np _snake_case = [8, 5, 9, 7] _snake_case = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] _snake_case = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class UpperCamelCase : def __init__( self : List[Any] , UpperCAmelCase__ : list[int] , UpperCAmelCase__ : list[list[int]] , UpperCAmelCase__ : list[list[int]] , ) -> None: _a : List[str] = claim_vector _a : List[Any] = allocated_resources_table _a : Union[str, Any] = maximum_claim_table def _lowercase ( self : Tuple ) -> list[int]: return [ sum(p_item[i] for p_item in self.__allocated_resources_table ) for i in range(len(self.__allocated_resources_table[0] ) ) ] def _lowercase ( self : int ) -> list[int]: return np.array(self.__claim_vector ) - np.array( self.__processes_resource_summation() ) def _lowercase ( self : List[str] ) -> list[list[int]]: return [ list(np.array(self.__maximum_claim_table[i] ) - np.array(UpperCAmelCase__ ) ) for i, allocated_resource in enumerate(self.__allocated_resources_table ) ] def _lowercase ( self : Optional[Any] ) -> dict[int, list[int]]: return {self.__need().index(UpperCAmelCase__ ): i for i in self.__need()} def _lowercase ( self : Dict , **UpperCAmelCase__ : Optional[Any] ) -> None: _a : List[Any] = self.__need() _a : Optional[int] = self.__allocated_resources_table _a : str = self.__available_resources() _a : Optional[Any] = self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print("""_""" * 50 + """\n""" ) while need_list: _a : int = False for each_need in need_list: _a : Optional[int] = True for index, need in enumerate(UpperCAmelCase__ ): if need > available_resources[index]: _a : List[Any] = False break if execution: _a : str = True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: _a : Any = original_need_index print(f"""Process {process_number + 1} is executing.""" ) # remove the process run from stack need_list.remove(UpperCAmelCase__ ) # update available/freed resources stack _a : Union[str, Any] = np.array(UpperCAmelCase__ ) + np.array( alloc_resources_table[process_number] ) print( """Updated available resource stack for processes: """ + """ """.join([str(UpperCAmelCase__ ) for x in available_resources] ) ) break if safe: print("""The process is in a safe state.\n""" ) else: print("""System in unsafe state. Aborting...\n""" ) break def _lowercase ( self : Any ) -> Optional[int]: print(""" """ * 9 + """Allocated Resource Table""" ) for item in self.__allocated_resources_table: print( f"""P{self.__allocated_resources_table.index(UpperCAmelCase__ ) + 1}""" + """ """.join(f"""{it:>8}""" for it in item ) + """\n""" ) print(""" """ * 9 + """System Resource Table""" ) for item in self.__maximum_claim_table: print( f"""P{self.__maximum_claim_table.index(UpperCAmelCase__ ) + 1}""" + """ """.join(f"""{it:>8}""" for it in item ) + """\n""" ) print( """Current Usage by Active Processes: """ + """ """.join(str(UpperCAmelCase__ ) for x in self.__claim_vector ) ) print( """Initial Available Resources: """ + """ """.join(str(UpperCAmelCase__ ) for x in self.__available_resources() ) ) time.sleep(1 ) if __name__ == "__main__": import doctest doctest.testmod()
294
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _snake_case = { 'configuration_graphormer': ['GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GraphormerConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'GraphormerForGraphClassification', 'GraphormerModel', 'GraphormerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_graphormer import GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, GraphormerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_graphormer import ( GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST, GraphormerForGraphClassification, GraphormerModel, GraphormerPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
294
"""simple docstring""" from collections.abc import Iterable from typing import Generic, TypeVar _snake_case = TypeVar('_T') class UpperCamelCase ( Generic[_T] ): def __init__( self : Optional[int] , UpperCAmelCase__ : Iterable[_T] | None = None ) -> None: _a : list[_T] = list(iterable or [] ) _a : list[_T] = [] def __len__( self : str ) -> int: return len(self._stacka ) + len(self._stacka ) def __repr__( self : List[str] ) -> str: return f"""Queue({tuple(self._stacka[::-1] + self._stacka )})""" def _lowercase ( self : Union[str, Any] , UpperCAmelCase__ : _T ) -> None: self._stacka.append(UpperCAmelCase__ ) def _lowercase ( self : Optional[Any] ) -> _T: _a : Any = self._stacka.pop _a : Union[str, Any] = self._stacka.append if not self._stacka: while self._stacka: stacka_append(stacka_pop() ) if not self._stacka: raise IndexError("""Queue is empty""" ) return self._stacka.pop() if __name__ == "__main__": from doctest import testmod testmod()
294
1
"""simple docstring""" import math def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' if 0 not in (x, y): # We use the relation x^y = y*log10(x), where 10 is the base. return y * math.logaa(UpperCamelCase__ ) else: if x == 0: # 0 raised to any number is 0 return 0 elif y == 0: return 1 # any number raised to 0 is 1 raise AssertionError("""This should never happen""" ) if __name__ == "__main__": # Main function # Read two numbers from input and typecast them to int using map function. # Here x is the base and y is the power. _snake_case = 'Enter the base and the power separated by a comma: ' _snake_case , _snake_case = map(int, input(prompt).split(',')) _snake_case , _snake_case = map(int, input(prompt).split(',')) # We find the log of each number, using the function res(), which takes two # arguments. _snake_case = res(xa, ya) _snake_case = res(xa, ya) # We check for the largest number if resa > resa: print('Largest number is', xa, '^', ya) elif resa > resa: print('Largest number is', xa, '^', ya) else: print('Both are equal')
294
"""simple docstring""" import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device _snake_case = False class UpperCamelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : Optional[Any] ) -> Optional[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self : Tuple ) -> List[Any]: _a : Tuple = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : List[str] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) _a : Optional[Any] = torch.manual_seed(0 ) _a : Union[str, Any] = pipe.dual_guided( prompt="""first prompt""" , image=UpperCAmelCase__ , text_to_image_strength=0.7_5 , generator=UpperCAmelCase__ , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(UpperCAmelCase__ ) _a : Dict = VersatileDiffusionPipeline.from_pretrained(UpperCAmelCase__ , torch_dtype=torch.floataa ) pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : Optional[Any] = generator.manual_seed(0 ) _a : str = pipe.dual_guided( prompt="""first prompt""" , image=UpperCAmelCase__ , text_to_image_strength=0.7_5 , generator=UpperCAmelCase__ , guidance_scale=7.5 , num_inference_steps=2 , output_type="""numpy""" , ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def _lowercase ( self : Optional[int] ) -> Optional[int]: _a : Optional[int] = VersatileDiffusionPipeline.from_pretrained("""shi-labs/versatile-diffusion""" , torch_dtype=torch.floataa ) pipe.to(UpperCAmelCase__ ) pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) _a : int = """cyberpunk 2077""" _a : Tuple = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) _a : Tuple = torch.manual_seed(0 ) _a : Any = pipe.dual_guided( prompt=UpperCAmelCase__ , image=UpperCAmelCase__ , text_to_image_strength=0.7_5 , generator=UpperCAmelCase__ , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" , ).images _a : List[str] = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _a : Optional[int] = np.array([0.1_4_4_8, 0.1_6_1_9, 0.1_7_4_1, 0.1_0_8_6, 0.1_1_4_7, 0.1_1_2_8, 0.1_1_9_9, 0.1_1_6_5, 0.1_0_0_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 _a : int = """A painting of a squirrel eating a burger """ _a : Tuple = torch.manual_seed(0 ) _a : Union[str, Any] = pipe.text_to_image( prompt=UpperCAmelCase__ , generator=UpperCAmelCase__ , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" ).images _a : int = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _a : int = np.array([0.3_3_6_7, 0.3_1_6_9, 0.2_6_5_6, 0.3_8_7_0, 0.4_7_9_0, 0.3_7_9_6, 0.4_0_0_9, 0.4_8_7_8, 0.4_7_7_8] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 _a : str = pipe.image_variation(UpperCAmelCase__ , generator=UpperCAmelCase__ , output_type="""numpy""" ).images _a : str = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) _a : Optional[Any] = np.array([0.3_0_7_6, 0.3_1_2_3, 0.3_2_8_4, 0.3_7_8_2, 0.3_7_7_0, 0.3_8_9_4, 0.4_2_9_7, 0.4_3_3_1, 0.4_4_5_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1
294
1
"""simple docstring""" def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' return int((input_a, input_a).count(0 ) != 0 ) def lowerCAmelCase__ ( ): '''simple docstring''' assert nand_gate(0 , 0 ) == 1 assert nand_gate(0 , 1 ) == 1 assert nand_gate(1 , 0 ) == 1 assert nand_gate(1 , 1 ) == 0 if __name__ == "__main__": print(nand_gate(0, 0)) print(nand_gate(0, 1)) print(nand_gate(1, 0)) print(nand_gate(1, 1))
294
"""simple docstring""" import json import os from typing import Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = {'vocab_file': 'vocab.json'} _snake_case = { 'vocab_file': { 'mgp-str': 'https://huggingface.co/alibaba-damo/mgp-str-base/blob/main/vocab.json', } } _snake_case = {'mgp-str': 27} class UpperCamelCase ( snake_case_ ): UpperCamelCase : List[str] = VOCAB_FILES_NAMES UpperCamelCase : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : Optional[int] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : List[Any]="[GO]" , UpperCAmelCase__ : Tuple="[GO]" , UpperCAmelCase__ : Optional[int]="[s]" , UpperCAmelCase__ : int="[GO]" , **UpperCAmelCase__ : Dict ) -> int: super().__init__( unk_token=UpperCAmelCase__ , bos_token=UpperCAmelCase__ , eos_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , **UpperCAmelCase__ , ) with open(UpperCAmelCase__ , encoding="""utf-8""" ) as vocab_handle: _a : int = json.load(UpperCAmelCase__ ) _a : Optional[int] = {v: k for k, v in self.vocab.items()} @property def _lowercase ( self : Dict ) -> Union[str, Any]: return len(self.vocab ) def _lowercase ( self : Union[str, Any] ) -> str: return dict(self.vocab , **self.added_tokens_encoder ) def _lowercase ( self : Dict , UpperCAmelCase__ : str ) -> Union[str, Any]: _a : Tuple = [] for s in text: char_tokens.extend(UpperCAmelCase__ ) return char_tokens def _lowercase ( self : List[Any] , UpperCAmelCase__ : str ) -> Dict: return self.vocab.get(UpperCAmelCase__ , self.vocab.get(self.unk_token ) ) def _lowercase ( self : Optional[int] , UpperCAmelCase__ : Tuple ) -> List[Any]: return self.decoder.get(UpperCAmelCase__ ) def _lowercase ( self : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(UpperCAmelCase__ ): logger.error("""Vocabulary path ({}) should be a directory""".format(UpperCAmelCase__ ) ) return _a : Tuple = os.path.join( UpperCAmelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) with open(UpperCAmelCase__ , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.vocab , indent=2 , sort_keys=UpperCAmelCase__ , ensure_ascii=UpperCAmelCase__ ) + """\n""" ) return (vocab_file,)
294
1
"""simple docstring""" import itertools import string from collections.abc import Generator, Iterable def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Dict = iter(UpperCamelCase__ ) while True: _a : List[str] = tuple(itertools.islice(UpperCamelCase__ , UpperCamelCase__ ) ) if not chunk: return yield chunk def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : Union[str, Any] = """""".join([c.upper() for c in dirty if c in string.ascii_letters] ) _a : str = """""" if len(UpperCamelCase__ ) < 2: return dirty for i in range(len(UpperCamelCase__ ) - 1 ): clean += dirty[i] if dirty[i] == dirty[i + 1]: clean += "X" clean += dirty[-1] if len(UpperCamelCase__ ) & 1: clean += "X" return clean def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' # I and J are used interchangeably to allow # us to use a 5x5 table (25 letters) _a : List[Any] = """ABCDEFGHIKLMNOPQRSTUVWXYZ""" # we're using a list instead of a '2d' array because it makes the math # for setting up the table and doing the actual encoding/decoding simpler _a : Tuple = [] # copy key chars into the table if they are in `alphabet` ignoring duplicates for char in key.upper(): if char not in table and char in alphabet: table.append(UpperCamelCase__ ) # fill the rest of the table in with the remaining alphabet chars for char in alphabet: if char not in table: table.append(UpperCamelCase__ ) return table def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Union[str, Any] = generate_table(UpperCamelCase__ ) _a : List[Any] = prepare_input(UpperCamelCase__ ) _a : Tuple = """""" # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(UpperCamelCase__ , 2 ): _a , _a : Optional[Any] = divmod(table.index(UpperCamelCase__ ) , 5 ) _a , _a : List[str] = divmod(table.index(UpperCamelCase__ ) , 5 ) if rowa == rowa: ciphertext += table[rowa * 5 + (cola + 1) % 5] ciphertext += table[rowa * 5 + (cola + 1) % 5] elif cola == cola: ciphertext += table[((rowa + 1) % 5) * 5 + cola] ciphertext += table[((rowa + 1) % 5) * 5 + cola] else: # rectangle ciphertext += table[rowa * 5 + cola] ciphertext += table[rowa * 5 + cola] return ciphertext def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Tuple = generate_table(UpperCamelCase__ ) _a : List[str] = """""" # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(UpperCamelCase__ , 2 ): _a , _a : Optional[int] = divmod(table.index(UpperCamelCase__ ) , 5 ) _a , _a : List[Any] = divmod(table.index(UpperCamelCase__ ) , 5 ) if rowa == rowa: plaintext += table[rowa * 5 + (cola - 1) % 5] plaintext += table[rowa * 5 + (cola - 1) % 5] elif cola == cola: plaintext += table[((rowa - 1) % 5) * 5 + cola] plaintext += table[((rowa - 1) % 5) * 5 + cola] else: # rectangle plaintext += table[rowa * 5 + cola] plaintext += table[rowa * 5 + cola] return plaintext
294
"""simple docstring""" import tempfile import torch from diffusers import IPNDMScheduler from .test_schedulers import SchedulerCommonTest class UpperCamelCase ( snake_case_ ): UpperCamelCase : int = (IPNDMScheduler,) UpperCamelCase : int = (('''num_inference_steps''', 50),) def _lowercase ( self : Union[str, Any] , **UpperCAmelCase__ : Tuple ) -> int: _a : Optional[int] = {"""num_train_timesteps""": 1000} config.update(**UpperCAmelCase__ ) return config def _lowercase ( self : Dict , UpperCAmelCase__ : Any=0 , **UpperCAmelCase__ : Optional[Any] ) -> Union[str, Any]: _a : Optional[int] = dict(self.forward_default_kwargs ) _a : Dict = kwargs.pop("""num_inference_steps""" , UpperCAmelCase__ ) _a : Optional[Any] = self.dummy_sample _a : Union[str, Any] = 0.1 * sample _a : Union[str, Any] = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] for scheduler_class in self.scheduler_classes: _a : Optional[int] = self.get_scheduler_config(**UpperCAmelCase__ ) _a : Union[str, Any] = scheduler_class(**UpperCAmelCase__ ) scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residuals _a : Any = dummy_past_residuals[:] if time_step is None: _a : str = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(UpperCAmelCase__ ) _a : Union[str, Any] = scheduler_class.from_pretrained(UpperCAmelCase__ ) new_scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residuals _a : Optional[Any] = dummy_past_residuals[:] _a : List[Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : str = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" _a : Optional[int] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : Tuple = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def _lowercase ( self : Tuple ) -> List[str]: pass def _lowercase ( self : Optional[int] , UpperCAmelCase__ : List[str]=0 , **UpperCAmelCase__ : Optional[Any] ) -> List[Any]: _a : Optional[Any] = dict(self.forward_default_kwargs ) _a : Optional[Any] = kwargs.pop("""num_inference_steps""" , UpperCAmelCase__ ) _a : Optional[Any] = self.dummy_sample _a : List[Any] = 0.1 * sample _a : Optional[Any] = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] for scheduler_class in self.scheduler_classes: _a : Union[str, Any] = self.get_scheduler_config() _a : Optional[Any] = scheduler_class(**UpperCAmelCase__ ) scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residuals (must be after setting timesteps) _a : Any = dummy_past_residuals[:] if time_step is None: _a : List[Any] = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(UpperCAmelCase__ ) _a : Any = scheduler_class.from_pretrained(UpperCAmelCase__ ) # copy over dummy past residuals new_scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residual (must be after setting timesteps) _a : Optional[Any] = dummy_past_residuals[:] _a : List[str] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : Tuple = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" _a : Union[str, Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : int = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def _lowercase ( self : str , **UpperCAmelCase__ : Any ) -> List[str]: _a : Optional[int] = self.scheduler_classes[0] _a : Optional[Any] = self.get_scheduler_config(**UpperCAmelCase__ ) _a : Union[str, Any] = scheduler_class(**UpperCAmelCase__ ) _a : int = 10 _a : List[Any] = self.dummy_model() _a : str = self.dummy_sample_deter scheduler.set_timesteps(UpperCAmelCase__ ) for i, t in enumerate(scheduler.timesteps ): _a : str = model(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ).prev_sample for i, t in enumerate(scheduler.timesteps ): _a : Union[str, Any] = model(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Any = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ).prev_sample return sample def _lowercase ( self : int ) -> str: _a : Dict = dict(self.forward_default_kwargs ) _a : int = kwargs.pop("""num_inference_steps""" , UpperCAmelCase__ ) for scheduler_class in self.scheduler_classes: _a : Optional[int] = self.get_scheduler_config() _a : Tuple = scheduler_class(**UpperCAmelCase__ ) _a : Tuple = self.dummy_sample _a : Optional[Any] = 0.1 * sample if num_inference_steps is not None and hasattr(UpperCAmelCase__ , """set_timesteps""" ): scheduler.set_timesteps(UpperCAmelCase__ ) elif num_inference_steps is not None and not hasattr(UpperCAmelCase__ , """set_timesteps""" ): _a : List[str] = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) _a : Union[str, Any] = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] _a : Optional[Any] = dummy_past_residuals[:] _a : Optional[Any] = scheduler.timesteps[5] _a : str = scheduler.timesteps[6] _a : Optional[int] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : Union[str, Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) _a : Tuple = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : List[str] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def _lowercase ( self : List[str] ) -> List[str]: for timesteps in [100, 1000]: self.check_over_configs(num_train_timesteps=UpperCAmelCase__ , time_step=UpperCAmelCase__ ) def _lowercase ( self : List[str] ) -> List[str]: for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 100] ): self.check_over_forward(num_inference_steps=UpperCAmelCase__ , time_step=UpperCAmelCase__ ) def _lowercase ( self : int ) -> List[Any]: _a : str = self.full_loop() _a : List[Any] = torch.mean(torch.abs(UpperCAmelCase__ ) ) assert abs(result_mean.item() - 2540529 ) < 10
294
1
"""simple docstring""" from __future__ import annotations def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' # preprocessing the first row for i in range(1 , len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1 , len(UpperCamelCase__ ) ): matrix[i][0] += matrix[i - 1][0] # updating the path cost for current position for i in range(1 , len(UpperCamelCase__ ) ): for j in range(1 , len(matrix[0] ) ): matrix[i][j] += min(matrix[i - 1][j] , matrix[i][j - 1] ) return matrix[-1][-1] if __name__ == "__main__": import doctest doctest.testmod()
294
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging _snake_case = logging.get_logger(__name__) _snake_case = {'vocab_file': 'spiece.model'} _snake_case = { 'vocab_file': { 'TsinghuaAI/CPM-Generate': 'https://huggingface.co/TsinghuaAI/CPM-Generate/resolve/main/spiece.model', } } class UpperCamelCase ( snake_case_ ): def __init__( self : Tuple , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : int=False , UpperCAmelCase__ : Dict=True , UpperCAmelCase__ : Union[str, Any]=False , UpperCAmelCase__ : Dict="<s>" , UpperCAmelCase__ : Any="</s>" , UpperCAmelCase__ : Any="<unk>" , UpperCAmelCase__ : int="<sep>" , UpperCAmelCase__ : Tuple="<pad>" , UpperCAmelCase__ : Any="<cls>" , UpperCAmelCase__ : Optional[Any]="<mask>" , UpperCAmelCase__ : int=["<eop>", "<eod>"] , UpperCAmelCase__ : Optional[Dict[str, Any]] = None , **UpperCAmelCase__ : List[str] , ) -> None: _a : Optional[int] = AddedToken(UpperCAmelCase__ , lstrip=UpperCAmelCase__ , rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else mask_token _a : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=UpperCAmelCase__ , remove_space=UpperCAmelCase__ , keep_accents=UpperCAmelCase__ , bos_token=UpperCAmelCase__ , eos_token=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , sep_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , cls_token=UpperCAmelCase__ , mask_token=UpperCAmelCase__ , additional_special_tokens=UpperCAmelCase__ , sp_model_kwargs=self.sp_model_kwargs , **UpperCAmelCase__ , ) _a : Optional[Any] = 3 _a : Tuple = do_lower_case _a : Tuple = remove_space _a : Tuple = keep_accents _a : Tuple = vocab_file _a : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCAmelCase__ ) try: import jieba except ModuleNotFoundError as error: raise error.__class__( """You need to install jieba to use CpmTokenizer or CpmTokenizerFast. """ """See https://pypi.org/project/jieba/ for installation.""" ) _a : int = jieba _a : Tuple = str.maketrans(""" \n""" , """\u2582\u2583""" ) @property # Copied from transformers.models.xlnet.tokenization_xlnet.XLNetTokenizer.vocab_size def _lowercase ( self : Optional[Any] ) -> Any: return len(self.sp_model ) def _lowercase ( self : str ) -> Union[str, Any]: _a : int = {self.convert_ids_to_tokens(UpperCAmelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Tuple ) -> List[str]: _a : Tuple = self.__dict__.copy() _a : Tuple = None return state def __setstate__( self : Any , UpperCAmelCase__ : Dict ) -> Dict: _a : Tuple = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): _a : Tuple = {} _a : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _lowercase ( self : List[str] , UpperCAmelCase__ : Union[str, Any] ) -> Dict: if self.remove_space: _a : Optional[int] = """ """.join(inputs.strip().split() ) else: _a : List[Any] = inputs _a : int = outputs.replace("""``""" , """\"""" ).replace("""''""" , """\"""" ) if not self.keep_accents: _a : Optional[Any] = unicodedata.normalize("""NFKD""" , UpperCAmelCase__ ) _a : Dict = """""".join([c for c in outputs if not unicodedata.combining(UpperCAmelCase__ )] ) if self.do_lower_case: _a : Union[str, Any] = outputs.lower() return outputs def _lowercase ( self : List[str] , UpperCAmelCase__ : str ) -> List[str]: _a : str = self.preprocess_text(UpperCAmelCase__ ) _a : Dict = self.sp_model.encode(UpperCAmelCase__ , out_type=UpperCAmelCase__ ) _a : Union[str, Any] = [] for piece in pieces: if len(UpperCAmelCase__ ) > 1 and piece[-1] == str(""",""" ) and piece[-2].isdigit(): _a : Dict = self.sp_model.EncodeAsPieces(piece[:-1].replace(UpperCAmelCase__ , """""" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: _a : Dict = cur_pieces[1:] else: _a : Any = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(UpperCAmelCase__ ) else: new_pieces.append(UpperCAmelCase__ ) return new_pieces def _lowercase ( self : Union[str, Any] , UpperCAmelCase__ : int ) -> int: return self.sp_model.PieceToId(UpperCAmelCase__ ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : Optional[Any] ) -> Any: return self.sp_model.IdToPiece(UpperCAmelCase__ ) def _lowercase ( self : Any , UpperCAmelCase__ : Any ) -> Dict: _a : Dict = """""".join(UpperCAmelCase__ ).replace(UpperCAmelCase__ , """ """ ).strip() return out_string def _lowercase ( self : Any , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ) -> List[int]: _a : Optional[Any] = [self.sep_token_id] _a : Dict = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _lowercase ( self : Tuple , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None , UpperCAmelCase__ : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase__ , token_ids_a=UpperCAmelCase__ , already_has_special_tokens=UpperCAmelCase__ ) if token_ids_a is not None: return ([0] * len(UpperCAmelCase__ )) + [1] + ([0] * len(UpperCAmelCase__ )) + [1, 1] return ([0] * len(UpperCAmelCase__ )) + [1, 1] def _lowercase ( self : str , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ) -> List[int]: _a : Any = [self.sep_token_id] _a : Optional[Any] = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _lowercase ( self : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(UpperCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return _a : Union[str, Any] = os.path.join( UpperCAmelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCAmelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCAmelCase__ , """wb""" ) as fi: _a : Optional[Any] = self.sp_model.serialized_model_proto() fi.write(UpperCAmelCase__ ) return (out_vocab_file,) def _lowercase ( self : Any , *UpperCAmelCase__ : List[str] , **UpperCAmelCase__ : List[str] ) -> List[str]: _a : Tuple = super()._decode(*UpperCAmelCase__ , **UpperCAmelCase__ ) _a : Optional[Any] = text.replace(""" """ , """""" ).replace("""\u2582""" , """ """ ).replace("""\u2583""" , """\n""" ) return text
294
1
"""simple docstring""" import unittest from typing import Tuple import torch from diffusers.utils import floats_tensor, randn_tensor, torch_all_close, torch_device from diffusers.utils.testing_utils import require_torch @require_torch class UpperCamelCase : @property def _lowercase ( self : Dict ) -> int: return self.get_dummy_input() @property def _lowercase ( self : Any ) -> List[Any]: if self.block_type == "down": return (4, 32, 16, 16) elif self.block_type == "mid": return (4, 32, 32, 32) elif self.block_type == "up": return (4, 32, 64, 64) raise ValueError(f"""'{self.block_type}' is not a supported block_type. Set it to 'up', 'mid', or 'down'.""" ) def _lowercase ( self : Tuple , UpperCAmelCase__ : Any=True , UpperCAmelCase__ : Any=False , UpperCAmelCase__ : List[str]=False , UpperCAmelCase__ : List[str]=False , ) -> Any: _a : Optional[Any] = 4 _a : Optional[int] = 32 _a : Union[str, Any] = (32, 32) _a : List[Any] = torch.manual_seed(0 ) _a : Tuple = torch.device(UpperCAmelCase__ ) _a : Tuple = (batch_size, num_channels) + sizes _a : str = randn_tensor(UpperCAmelCase__ , generator=UpperCAmelCase__ , device=UpperCAmelCase__ ) _a : List[str] = {"""hidden_states""": hidden_states} if include_temb: _a : Union[str, Any] = 128 _a : int = randn_tensor((batch_size, temb_channels) , generator=UpperCAmelCase__ , device=UpperCAmelCase__ ) if include_res_hidden_states_tuple: _a : Any = torch.manual_seed(1 ) _a : int = (randn_tensor(UpperCAmelCase__ , generator=UpperCAmelCase__ , device=UpperCAmelCase__ ),) if include_encoder_hidden_states: _a : Tuple = floats_tensor((batch_size, 32, 32) ).to(UpperCAmelCase__ ) if include_skip_sample: _a : int = randn_tensor(((batch_size, 3) + sizes) , generator=UpperCAmelCase__ , device=UpperCAmelCase__ ) return dummy_input def _lowercase ( self : Optional[int] ) -> Union[str, Any]: _a : Any = { """in_channels""": 32, """out_channels""": 32, """temb_channels""": 128, } if self.block_type == "up": _a : Optional[int] = 32 if self.block_type == "mid": init_dict.pop("""out_channels""" ) _a : Any = self.dummy_input return init_dict, inputs_dict def _lowercase ( self : str , UpperCAmelCase__ : Any ) -> Optional[int]: _a , _a : Any = self.prepare_init_args_and_inputs_for_common() _a : List[str] = self.block_class(**UpperCAmelCase__ ) unet_block.to(UpperCAmelCase__ ) unet_block.eval() with torch.no_grad(): _a : Tuple = unet_block(**UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): _a : List[Any] = output[0] self.assertEqual(output.shape , self.output_shape ) _a : Dict = output[0, -1, -3:, -3:] _a : Union[str, Any] = torch.tensor(UpperCAmelCase__ ).to(UpperCAmelCase__ ) assert torch_all_close(output_slice.flatten() , UpperCAmelCase__ , atol=5E-3 ) @unittest.skipIf(torch_device == """mps""" , """Training is not supported in mps""" ) def _lowercase ( self : Dict ) -> Optional[Any]: _a , _a : Any = self.prepare_init_args_and_inputs_for_common() _a : List[Any] = self.block_class(**UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.train() _a : List[str] = model(**UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): _a : Union[str, Any] = output[0] _a : Any = torch.device(UpperCAmelCase__ ) _a : Union[str, Any] = randn_tensor(output.shape , device=UpperCAmelCase__ ) _a : str = torch.nn.functional.mse_loss(UpperCAmelCase__ , UpperCAmelCase__ ) loss.backward()
294
"""simple docstring""" import inspect import unittest import torch import torch.nn as nn from accelerate.hooks import ( AlignDevicesHook, ModelHook, SequentialHook, add_hook_to_module, attach_align_device_hook, remove_hook_from_module, remove_hook_from_submodules, ) from accelerate.test_utils import require_multi_gpu class UpperCamelCase ( nn.Module ): def __init__( self : Union[str, Any] ) -> int: super().__init__() _a : Optional[Any] = nn.Linear(3 , 4 ) _a : Tuple = nn.BatchNormad(4 ) _a : Dict = nn.Linear(4 , 5 ) def _lowercase ( self : Optional[int] , UpperCAmelCase__ : List[str] ) -> int: return self.lineara(self.batchnorm(self.lineara(UpperCAmelCase__ ) ) ) class UpperCamelCase ( snake_case_ ): def _lowercase ( self : Any , UpperCAmelCase__ : Any , *UpperCAmelCase__ : List[str] , **UpperCAmelCase__ : Optional[int] ) -> Optional[int]: return (args[0] + 1,) + args[1:], kwargs class UpperCamelCase ( snake_case_ ): def _lowercase ( self : Optional[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Optional[Any] ) -> List[str]: return output + 1 class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : Dict ) -> str: _a : List[Any] = ModelForTest() _a : str = ModelHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) self.assertEqual(test_model._hf_hook , UpperCAmelCase__ ) self.assertTrue(hasattr(UpperCAmelCase__ , """_old_forward""" ) ) # Check adding the hook did not change the name or the signature self.assertEqual(test_model.forward.__name__ , """forward""" ) self.assertListEqual(list(inspect.signature(test_model.forward ).parameters ) , ["""x"""] ) remove_hook_from_module(UpperCAmelCase__ ) self.assertFalse(hasattr(UpperCAmelCase__ , """_hf_hook""" ) ) self.assertFalse(hasattr(UpperCAmelCase__ , """_old_forward""" ) ) def _lowercase ( self : Optional[int] ) -> Optional[int]: _a : Dict = ModelForTest() _a : Dict = ModelHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ , append=UpperCAmelCase__ ) self.assertEqual(isinstance(test_model._hf_hook , UpperCAmelCase__ ) , UpperCAmelCase__ ) self.assertEqual(len(test_model._hf_hook.hooks ) , 2 ) self.assertTrue(hasattr(UpperCAmelCase__ , """_old_forward""" ) ) # Check adding the hook did not change the name or the signature self.assertEqual(test_model.forward.__name__ , """forward""" ) self.assertListEqual(list(inspect.signature(test_model.forward ).parameters ) , ["""x"""] ) remove_hook_from_module(UpperCAmelCase__ ) self.assertFalse(hasattr(UpperCAmelCase__ , """_hf_hook""" ) ) self.assertFalse(hasattr(UpperCAmelCase__ , """_old_forward""" ) ) def _lowercase ( self : Dict ) -> int: _a : str = ModelForTest() _a : List[Any] = torch.randn(2 , 3 ) _a : Optional[Any] = test_model(x + 1 ) _a : str = test_model(x + 2 ) _a : Union[str, Any] = PreForwardHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Tuple = test_model(UpperCAmelCase__ ) self.assertTrue(torch.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1E-5 ) ) # Attaching a hook to a model when it already has one replaces, does not chain _a : int = PreForwardHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : str = test_model(UpperCAmelCase__ ) self.assertTrue(torch.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1E-5 ) ) # You need to use the sequential hook to chain two or more hooks _a : int = SequentialHook(PreForwardHook() , PreForwardHook() ) add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Tuple = test_model(UpperCAmelCase__ ) assert torch.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1E-5 ) def _lowercase ( self : Tuple ) -> int: _a : Tuple = ModelForTest() _a : Union[str, Any] = torch.randn(2 , 3 ) _a : Optional[int] = test_model(UpperCAmelCase__ ) _a : int = PostForwardHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[str] = test_model(UpperCAmelCase__ ) self.assertTrue(torch.allclose(UpperCAmelCase__ , output + 1 , atol=1E-5 ) ) # Attaching a hook to a model when it already has one replaces, does not chain _a : List[Any] = PostForwardHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Dict = test_model(UpperCAmelCase__ ) self.assertTrue(torch.allclose(UpperCAmelCase__ , output + 1 , atol=1E-5 ) ) # You need to use the sequential hook to chain two or more hooks _a : Any = SequentialHook(PostForwardHook() , PostForwardHook() ) add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Optional[int] = test_model(UpperCAmelCase__ ) assert torch.allclose(UpperCAmelCase__ , output + 2 , atol=1E-5 ) def _lowercase ( self : Dict ) -> Optional[Any]: _a : Any = ModelForTest() _a : List[Any] = torch.randn(2 , 3 ) _a : Dict = test_model(UpperCAmelCase__ ) _a : Any = PostForwardHook() add_hook_to_module(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[str] = test_model(UpperCAmelCase__ ) self.assertTrue(torch.allclose(UpperCAmelCase__ , output + 1 ) ) self.assertTrue(outputa.requires_grad ) _a : Any = True _a : Union[str, Any] = test_model(UpperCAmelCase__ ) self.assertFalse(outputa.requires_grad ) @require_multi_gpu def _lowercase ( self : Optional[Any] ) -> str: _a : List[Any] = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices add_hook_to_module(model.lineara , AlignDevicesHook(execution_device=0 ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(execution_device=0 ) ) add_hook_to_module(model.lineara , AlignDevicesHook(execution_device=1 ) ) self.assertEqual(model.lineara.weight.device , torch.device(0 ) ) self.assertEqual(model.batchnorm.weight.device , torch.device(0 ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device(0 ) ) self.assertEqual(model.lineara.weight.device , torch.device(1 ) ) # We can still make a forward pass. The input does not need to be on any particular device _a : Optional[int] = torch.randn(2 , 3 ) _a : Any = model(UpperCAmelCase__ ) self.assertEqual(output.device , torch.device(1 ) ) # We can add a general hook to put back output on same device as input. add_hook_to_module(UpperCAmelCase__ , AlignDevicesHook(io_same_device=UpperCAmelCase__ ) ) _a : str = torch.randn(2 , 3 ).to(0 ) _a : Union[str, Any] = model(UpperCAmelCase__ ) self.assertEqual(output.device , torch.device(0 ) ) def _lowercase ( self : str ) -> Union[str, Any]: _a : int = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices _a : List[Any] = {"""execution_device""": 0 if torch.cuda.is_available() else """cpu""", """offload""": True} add_hook_to_module(model.lineara , AlignDevicesHook(**UpperCAmelCase__ ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(**UpperCAmelCase__ ) ) add_hook_to_module(model.lineara , AlignDevicesHook(**UpperCAmelCase__ ) ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) # Buffers are not included in the offload by default, so are on the execution device _a : Dict = torch.device(hook_kwargs["""execution_device"""] ) self.assertEqual(model.batchnorm.running_mean.device , UpperCAmelCase__ ) _a : int = torch.randn(2 , 3 ) _a : str = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_module(model.lineara ) remove_hook_from_module(model.batchnorm ) remove_hook_from_module(model.lineara ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # Now test with buffers included in the offload _a : List[str] = { """execution_device""": 0 if torch.cuda.is_available() else """cpu""", """offload""": True, """offload_buffers""": True, } add_hook_to_module(model.lineara , AlignDevicesHook(**UpperCAmelCase__ ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(**UpperCAmelCase__ ) ) add_hook_to_module(model.lineara , AlignDevicesHook(**UpperCAmelCase__ ) ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device("""meta""" ) ) _a : Tuple = torch.randn(2 , 3 ) _a : Union[str, Any] = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_module(model.lineara ) remove_hook_from_module(model.batchnorm ) remove_hook_from_module(model.lineara ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) def _lowercase ( self : Tuple ) -> List[str]: _a : str = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices _a : Union[str, Any] = 0 if torch.cuda.is_available() else """cpu""" attach_align_device_hook(UpperCAmelCase__ , execution_device=UpperCAmelCase__ , offload=UpperCAmelCase__ ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) # Buffers are not included in the offload by default, so are on the execution device _a : Dict = torch.device(UpperCAmelCase__ ) self.assertEqual(model.batchnorm.running_mean.device , UpperCAmelCase__ ) _a : Union[str, Any] = torch.randn(2 , 3 ) _a : List[Any] = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(UpperCAmelCase__ ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # Now test with buffers included in the offload attach_align_device_hook(UpperCAmelCase__ , execution_device=UpperCAmelCase__ , offload=UpperCAmelCase__ , offload_buffers=UpperCAmelCase__ ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device("""meta""" ) ) _a : List[str] = torch.randn(2 , 3 ) _a : Union[str, Any] = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(UpperCAmelCase__ ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) def _lowercase ( self : Dict ) -> str: _a : Optional[Any] = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # This will move each submodule on different devices _a : str = 0 if torch.cuda.is_available() else """cpu""" attach_align_device_hook( UpperCAmelCase__ , execution_device=UpperCAmelCase__ , offload=UpperCAmelCase__ , weights_map=model.state_dict() ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) # Buffers are not included in the offload by default, so are on the execution device _a : Union[str, Any] = torch.device(UpperCAmelCase__ ) self.assertEqual(model.batchnorm.running_mean.device , UpperCAmelCase__ ) _a : Union[str, Any] = torch.randn(2 , 3 ) _a : int = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(UpperCAmelCase__ ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) # Now test with buffers included in the offload attach_align_device_hook( UpperCAmelCase__ , execution_device=UpperCAmelCase__ , offload=UpperCAmelCase__ , weights_map=model.state_dict() , offload_buffers=UpperCAmelCase__ , ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""meta""" ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device("""meta""" ) ) _a : Any = torch.randn(2 , 3 ) _a : int = model(UpperCAmelCase__ ) self.assertEqual(output.device , UpperCAmelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(UpperCAmelCase__ ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.batchnorm.weight.device , torch.device("""cpu""" ) ) self.assertEqual(model.lineara.weight.device , torch.device("""cpu""" ) )
294
1
"""simple docstring""" import unittest from transformers import LiltConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, ) from transformers.models.lilt.modeling_lilt import LILT_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCamelCase : def __init__( self : List[str] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : str=13 , UpperCAmelCase__ : Optional[Any]=7 , UpperCAmelCase__ : List[Any]=True , UpperCAmelCase__ : Optional[int]=True , UpperCAmelCase__ : Optional[Any]=True , UpperCAmelCase__ : Any=True , UpperCAmelCase__ : Any=99 , UpperCAmelCase__ : Tuple=24 , UpperCAmelCase__ : Any=2 , UpperCAmelCase__ : List[Any]=6 , UpperCAmelCase__ : Any=37 , UpperCAmelCase__ : Optional[int]="gelu" , UpperCAmelCase__ : int=0.1 , UpperCAmelCase__ : List[Any]=0.1 , UpperCAmelCase__ : Union[str, Any]=512 , UpperCAmelCase__ : Optional[int]=16 , UpperCAmelCase__ : Dict=2 , UpperCAmelCase__ : Optional[int]=0.0_2 , UpperCAmelCase__ : Union[str, Any]=3 , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : Dict=1000 , ) -> str: _a : Dict = parent _a : Dict = batch_size _a : Tuple = seq_length _a : List[str] = is_training _a : Optional[int] = use_input_mask _a : Any = use_token_type_ids _a : List[Any] = use_labels _a : Dict = vocab_size _a : Optional[Any] = hidden_size _a : Any = num_hidden_layers _a : Union[str, Any] = num_attention_heads _a : List[str] = intermediate_size _a : List[Any] = hidden_act _a : int = hidden_dropout_prob _a : List[Any] = attention_probs_dropout_prob _a : Optional[int] = max_position_embeddings _a : Union[str, Any] = type_vocab_size _a : int = type_sequence_label_size _a : Optional[Any] = initializer_range _a : Union[str, Any] = num_labels _a : List[Any] = scope _a : List[str] = range_bbox def _lowercase ( self : Tuple ) -> str: _a : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _a : Optional[Any] = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: _a : List[str] = bbox[i, j, 3] _a : List[Any] = bbox[i, j, 1] _a : List[Any] = t if bbox[i, j, 2] < bbox[i, j, 0]: _a : List[str] = bbox[i, j, 2] _a : Tuple = bbox[i, j, 0] _a : Tuple = t _a : Any = None if self.use_input_mask: _a : Tuple = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) _a : Any = None if self.use_token_type_ids: _a : str = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _a : Optional[Any] = None _a : Tuple = None if self.use_labels: _a : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _a : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _a : Tuple = self.get_config() return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels def _lowercase ( self : Optional[Any] ) -> Any: return LiltConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) def _lowercase ( self : int , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : int , ) -> Union[str, Any]: _a : List[str] = LiltModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() _a : List[str] = model(UpperCAmelCase__ , bbox=UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ ) _a : List[Any] = model(UpperCAmelCase__ , bbox=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ ) _a : Any = model(UpperCAmelCase__ , bbox=UpperCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _lowercase ( self : Tuple , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : str , UpperCAmelCase__ : List[str] , ) -> Optional[Any]: _a : str = self.num_labels _a : List[str] = LiltForTokenClassification(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() _a : Dict = model( UpperCAmelCase__ , bbox=UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowercase ( self : Union[str, Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Any , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Union[str, Any] , ) -> str: _a : List[str] = LiltForQuestionAnswering(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() _a : str = model( UpperCAmelCase__ , bbox=UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , token_type_ids=UpperCAmelCase__ , start_positions=UpperCAmelCase__ , end_positions=UpperCAmelCase__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowercase ( self : Optional[int] ) -> Optional[Any]: _a : List[str] = self.prepare_config_and_inputs() ( ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ) : Optional[Any] = config_and_inputs _a : str = { """input_ids""": input_ids, """bbox""": bbox, """token_type_ids""": token_type_ids, """attention_mask""": input_mask, } return config, inputs_dict @require_torch class UpperCamelCase ( snake_case_ , snake_case_ , snake_case_ , unittest.TestCase ): UpperCamelCase : List[Any] = ( ( LiltModel, LiltForSequenceClassification, LiltForTokenClassification, LiltForQuestionAnswering, ) if is_torch_available() else () ) UpperCamelCase : Tuple = ( { '''feature-extraction''': LiltModel, '''question-answering''': LiltForQuestionAnswering, '''text-classification''': LiltForSequenceClassification, '''token-classification''': LiltForTokenClassification, '''zero-shot''': LiltForSequenceClassification, } if is_torch_available() else {} ) UpperCamelCase : Dict = False UpperCamelCase : Union[str, Any] = False def _lowercase ( self : Dict , UpperCAmelCase__ : str , UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : int , UpperCAmelCase__ : Union[str, Any] ) -> Tuple: return True def _lowercase ( self : List[str] ) -> Optional[Any]: _a : Dict = LiltModelTester(self ) _a : Optional[Any] = ConfigTester(self , config_class=UpperCAmelCase__ , hidden_size=37 ) def _lowercase ( self : List[str] ) -> Tuple: self.config_tester.run_common_tests() def _lowercase ( self : List[Any] ) -> Union[str, Any]: _a : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase__ ) def _lowercase ( self : Dict ) -> Optional[int]: _a : Optional[int] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _a : str = type self.model_tester.create_and_check_model(*UpperCAmelCase__ ) def _lowercase ( self : Union[str, Any] ) -> str: _a : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*UpperCAmelCase__ ) def _lowercase ( self : Tuple ) -> Optional[int]: _a : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*UpperCAmelCase__ ) @slow def _lowercase ( self : Optional[Any] ) -> List[Any]: for model_name in LILT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _a : List[Any] = LiltModel.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) @require_torch @slow class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : int ) -> List[Any]: _a : Tuple = LiltModel.from_pretrained("""SCUT-DLVCLab/lilt-roberta-en-base""" ).to(UpperCAmelCase__ ) _a : Union[str, Any] = torch.tensor([[1, 2]] , device=UpperCAmelCase__ ) _a : List[Any] = torch.tensor([[[1, 2, 3, 4], [5, 6, 7, 8]]] , device=UpperCAmelCase__ ) # forward pass with torch.no_grad(): _a : Union[str, Any] = model(input_ids=UpperCAmelCase__ , bbox=UpperCAmelCase__ ) _a : str = torch.Size([1, 2, 768] ) _a : Any = torch.tensor( [[-0.0_6_5_3, 0.0_9_5_0, -0.0_0_6_1], [-0.0_5_4_5, 0.0_9_2_6, -0.0_3_2_4]] , device=UpperCAmelCase__ , ) self.assertTrue(outputs.last_hidden_state.shape , UpperCAmelCase__ ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :, :3] , UpperCAmelCase__ , atol=1E-3 ) )
294
"""simple docstring""" from __future__ import annotations def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' print(F"""Vertex\tShortest Distance from vertex {src}""" ) for i, d in enumerate(UpperCamelCase__ ): print(F"""{i}\t\t{d}""" ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' for j in range(UpperCamelCase__ ): _a , _a , _a : List[str] = (graph[j][k] for k in ["""src""", """dst""", """weight"""]) if distance[u] != float("""inf""" ) and distance[u] + w < distance[v]: return True return False def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Dict = [float("""inf""" )] * vertex_count _a : Any = 0.0 for _ in range(vertex_count - 1 ): for j in range(UpperCamelCase__ ): _a , _a , _a : List[Any] = (graph[j][k] for k in ["""src""", """dst""", """weight"""]) if distance[u] != float("""inf""" ) and distance[u] + w < distance[v]: _a : Any = distance[u] + w _a : Union[str, Any] = check_negative_cycle(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) if negative_cycle_exists: raise Exception("""Negative cycle found""" ) return distance if __name__ == "__main__": import doctest doctest.testmod() _snake_case = int(input('Enter number of vertices: ').strip()) _snake_case = int(input('Enter number of edges: ').strip()) _snake_case = [{} for _ in range(E)] for i in range(E): print('Edge ', i + 1) _snake_case , _snake_case , _snake_case = ( int(x) for x in input('Enter source, destination, weight: ').strip().split(' ') ) _snake_case = {'src': src, 'dst': dest, 'weight': weight} _snake_case = int(input('\nEnter shortest path source:').strip()) _snake_case = bellman_ford(graph, V, E, source) print_distance(shortest_distance, 0)
294
1
"""simple docstring""" import tempfile import torch from diffusers import IPNDMScheduler from .test_schedulers import SchedulerCommonTest class UpperCamelCase ( snake_case_ ): UpperCamelCase : int = (IPNDMScheduler,) UpperCamelCase : int = (('''num_inference_steps''', 50),) def _lowercase ( self : Union[str, Any] , **UpperCAmelCase__ : Tuple ) -> int: _a : Optional[int] = {"""num_train_timesteps""": 1000} config.update(**UpperCAmelCase__ ) return config def _lowercase ( self : Dict , UpperCAmelCase__ : Any=0 , **UpperCAmelCase__ : Optional[Any] ) -> Union[str, Any]: _a : Optional[int] = dict(self.forward_default_kwargs ) _a : Dict = kwargs.pop("""num_inference_steps""" , UpperCAmelCase__ ) _a : Optional[Any] = self.dummy_sample _a : Union[str, Any] = 0.1 * sample _a : Union[str, Any] = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] for scheduler_class in self.scheduler_classes: _a : Optional[int] = self.get_scheduler_config(**UpperCAmelCase__ ) _a : Union[str, Any] = scheduler_class(**UpperCAmelCase__ ) scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residuals _a : Any = dummy_past_residuals[:] if time_step is None: _a : str = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(UpperCAmelCase__ ) _a : Union[str, Any] = scheduler_class.from_pretrained(UpperCAmelCase__ ) new_scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residuals _a : Optional[Any] = dummy_past_residuals[:] _a : List[Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : str = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" _a : Optional[int] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : Tuple = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def _lowercase ( self : Tuple ) -> List[str]: pass def _lowercase ( self : Optional[int] , UpperCAmelCase__ : List[str]=0 , **UpperCAmelCase__ : Optional[Any] ) -> List[Any]: _a : Optional[Any] = dict(self.forward_default_kwargs ) _a : Optional[Any] = kwargs.pop("""num_inference_steps""" , UpperCAmelCase__ ) _a : Optional[Any] = self.dummy_sample _a : List[Any] = 0.1 * sample _a : Optional[Any] = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] for scheduler_class in self.scheduler_classes: _a : Union[str, Any] = self.get_scheduler_config() _a : Optional[Any] = scheduler_class(**UpperCAmelCase__ ) scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residuals (must be after setting timesteps) _a : Any = dummy_past_residuals[:] if time_step is None: _a : List[Any] = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(UpperCAmelCase__ ) _a : Any = scheduler_class.from_pretrained(UpperCAmelCase__ ) # copy over dummy past residuals new_scheduler.set_timesteps(UpperCAmelCase__ ) # copy over dummy past residual (must be after setting timesteps) _a : Optional[Any] = dummy_past_residuals[:] _a : List[str] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : Tuple = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" _a : Union[str, Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : int = new_scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def _lowercase ( self : str , **UpperCAmelCase__ : Any ) -> List[str]: _a : Optional[int] = self.scheduler_classes[0] _a : Optional[Any] = self.get_scheduler_config(**UpperCAmelCase__ ) _a : Union[str, Any] = scheduler_class(**UpperCAmelCase__ ) _a : int = 10 _a : List[Any] = self.dummy_model() _a : str = self.dummy_sample_deter scheduler.set_timesteps(UpperCAmelCase__ ) for i, t in enumerate(scheduler.timesteps ): _a : str = model(UpperCAmelCase__ , UpperCAmelCase__ ) _a : List[Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ).prev_sample for i, t in enumerate(scheduler.timesteps ): _a : Union[str, Any] = model(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Any = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ).prev_sample return sample def _lowercase ( self : int ) -> str: _a : Dict = dict(self.forward_default_kwargs ) _a : int = kwargs.pop("""num_inference_steps""" , UpperCAmelCase__ ) for scheduler_class in self.scheduler_classes: _a : Optional[int] = self.get_scheduler_config() _a : Tuple = scheduler_class(**UpperCAmelCase__ ) _a : Tuple = self.dummy_sample _a : Optional[Any] = 0.1 * sample if num_inference_steps is not None and hasattr(UpperCAmelCase__ , """set_timesteps""" ): scheduler.set_timesteps(UpperCAmelCase__ ) elif num_inference_steps is not None and not hasattr(UpperCAmelCase__ , """set_timesteps""" ): _a : List[str] = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) _a : Union[str, Any] = [residual + 0.2, residual + 0.1_5, residual + 0.1, residual + 0.0_5] _a : Optional[Any] = dummy_past_residuals[:] _a : Optional[Any] = scheduler.timesteps[5] _a : str = scheduler.timesteps[6] _a : Optional[int] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : Union[str, Any] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) _a : Tuple = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample _a : List[str] = scheduler.step(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def _lowercase ( self : List[str] ) -> List[str]: for timesteps in [100, 1000]: self.check_over_configs(num_train_timesteps=UpperCAmelCase__ , time_step=UpperCAmelCase__ ) def _lowercase ( self : List[str] ) -> List[str]: for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 100] ): self.check_over_forward(num_inference_steps=UpperCAmelCase__ , time_step=UpperCAmelCase__ ) def _lowercase ( self : int ) -> List[Any]: _a : str = self.full_loop() _a : List[Any] = torch.mean(torch.abs(UpperCAmelCase__ ) ) assert abs(result_mean.item() - 2540529 ) < 10
294
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _snake_case = { 'configuration_transfo_xl': ['TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TransfoXLConfig'], 'tokenization_transfo_xl': ['TransfoXLCorpus', 'TransfoXLTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST', 'AdaptiveEmbedding', 'TransfoXLForSequenceClassification', 'TransfoXLLMHeadModel', 'TransfoXLModel', 'TransfoXLPreTrainedModel', 'load_tf_weights_in_transfo_xl', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFAdaptiveEmbedding', 'TFTransfoXLForSequenceClassification', 'TFTransfoXLLMHeadModel', 'TFTransfoXLMainLayer', 'TFTransfoXLModel', 'TFTransfoXLPreTrainedModel', ] if TYPE_CHECKING: from .configuration_transfo_xl import TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, TransfoXLConfig from .tokenization_transfo_xl import TransfoXLCorpus, TransfoXLTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_transfo_xl import ( TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, AdaptiveEmbedding, TransfoXLForSequenceClassification, TransfoXLLMHeadModel, TransfoXLModel, TransfoXLPreTrainedModel, load_tf_weights_in_transfo_xl, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_transfo_xl import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFAdaptiveEmbedding, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLMainLayer, TFTransfoXLModel, TFTransfoXLPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
294
1
"""simple docstring""" from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
294
"""simple docstring""" def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , ): '''simple docstring''' _a : Optional[Any] = [redshift, radiation_density, matter_density, dark_energy] if any(p < 0 for p in parameters ): raise ValueError("""All input parameters must be positive""" ) if any(p > 1 for p in parameters[1:4] ): raise ValueError("""Relative densities cannot be greater than one""" ) else: _a : Tuple = 1 - (matter_density + radiation_density + dark_energy) _a : int = ( radiation_density * (redshift + 1) ** 4 + matter_density * (redshift + 1) ** 3 + curvature * (redshift + 1) ** 2 + dark_energy ) _a : List[str] = hubble_constant * e_a ** (1 / 2) return hubble if __name__ == "__main__": import doctest # run doctest doctest.testmod() # demo LCDM approximation _snake_case = 0.3 print( hubble_parameter( hubble_constant=68.3, radiation_density=1e-4, matter_density=matter_density, dark_energy=1 - matter_density, redshift=0, ) )
294
1
"""simple docstring""" from ...processing_utils import ProcessorMixin class UpperCamelCase ( snake_case_ ): UpperCamelCase : Union[str, Any] = '''WhisperFeatureExtractor''' UpperCamelCase : List[Any] = '''WhisperTokenizer''' def __init__( self : str , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : List[str] ) -> str: super().__init__(UpperCAmelCase__ , UpperCAmelCase__ ) _a : Optional[int] = self.feature_extractor _a : Dict = False def _lowercase ( self : Tuple , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : int=None , UpperCAmelCase__ : Any=True ) -> Tuple: return self.tokenizer.get_decoder_prompt_ids(task=UpperCAmelCase__ , language=UpperCAmelCase__ , no_timestamps=UpperCAmelCase__ ) def __call__( self : Tuple , *UpperCAmelCase__ : Optional[int] , **UpperCAmelCase__ : List[Any] ) -> Union[str, Any]: # For backward compatibility if self._in_target_context_manager: return self.current_processor(*UpperCAmelCase__ , **UpperCAmelCase__ ) _a : Dict = kwargs.pop("""audio""" , UpperCAmelCase__ ) _a : Optional[Any] = kwargs.pop("""sampling_rate""" , UpperCAmelCase__ ) _a : List[Any] = kwargs.pop("""text""" , UpperCAmelCase__ ) if len(UpperCAmelCase__ ) > 0: _a : Optional[int] = args[0] _a : Optional[int] = args[1:] if audio is None and text is None: raise ValueError("""You need to specify either an `audio` or `text` input to process.""" ) if audio is not None: _a : List[Any] = self.feature_extractor(UpperCAmelCase__ , *UpperCAmelCase__ , sampling_rate=UpperCAmelCase__ , **UpperCAmelCase__ ) if text is not None: _a : List[str] = self.tokenizer(UpperCAmelCase__ , **UpperCAmelCase__ ) if text is None: return inputs elif audio is None: return encodings else: _a : Any = encodings["""input_ids"""] return inputs def _lowercase ( self : Optional[int] , *UpperCAmelCase__ : str , **UpperCAmelCase__ : List[Any] ) -> str: return self.tokenizer.batch_decode(*UpperCAmelCase__ , **UpperCAmelCase__ ) def _lowercase ( self : Dict , *UpperCAmelCase__ : Any , **UpperCAmelCase__ : int ) -> Union[str, Any]: return self.tokenizer.decode(*UpperCAmelCase__ , **UpperCAmelCase__ ) def _lowercase ( self : str , UpperCAmelCase__ : str , UpperCAmelCase__ : List[Any]="np" ) -> Any: return self.tokenizer.get_prompt_ids(UpperCAmelCase__ , return_tensors=UpperCAmelCase__ )
294
"""simple docstring""" import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : List[Any] ) -> Dict: _a : Optional[int] = tempfile.mkdtemp() _a : Optional[Any] = SamImageProcessor() _a : int = SamProcessor(UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) def _lowercase ( self : Tuple , **UpperCAmelCase__ : Any ) -> Any: return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase__ ).image_processor def _lowercase ( self : str ) -> int: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Tuple ) -> Dict: _a : List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _a : Optional[int] = [Image.fromarray(np.moveaxis(UpperCAmelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def _lowercase ( self : Dict ) -> Dict: _a : List[Any] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _a : Tuple = self.get_image_processor(do_normalize=UpperCAmelCase__ , padding_value=1.0 ) _a : Tuple = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=UpperCAmelCase__ , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCAmelCase__ ) def _lowercase ( self : Union[str, Any] ) -> Tuple: _a : Optional[Any] = self.get_image_processor() _a : int = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Union[str, Any] = self.prepare_image_inputs() _a : List[str] = image_processor(UpperCAmelCase__ , return_tensors="""np""" ) _a : List[str] = processor(images=UpperCAmelCase__ , return_tensors="""np""" ) input_feat_extract.pop("""original_sizes""" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("""reshaped_input_sizes""" ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_torch def _lowercase ( self : Optional[Any] ) -> Optional[Any]: _a : Optional[Any] = self.get_image_processor() _a : Dict = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Tuple = [torch.ones((1, 3, 5, 5) )] _a : Tuple = [[1764, 2646]] _a : Optional[int] = [[683, 1024]] _a : List[Any] = processor.post_process_masks(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) _a : int = processor.post_process_masks( UpperCAmelCase__ , torch.tensor(UpperCAmelCase__ ) , torch.tensor(UpperCAmelCase__ ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np _a : Optional[Any] = [np.ones((1, 3, 5, 5) )] _a : Tuple = processor.post_process_masks(UpperCAmelCase__ , np.array(UpperCAmelCase__ ) , np.array(UpperCAmelCase__ ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) _a : List[str] = [[1, 0], [0, 1]] with self.assertRaises(UpperCAmelCase__ ): _a : str = processor.post_process_masks(UpperCAmelCase__ , np.array(UpperCAmelCase__ ) , np.array(UpperCAmelCase__ ) ) @require_vision @require_tf class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : Any ) -> List[str]: _a : List[str] = tempfile.mkdtemp() _a : Any = SamImageProcessor() _a : Union[str, Any] = SamProcessor(UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) def _lowercase ( self : List[str] , **UpperCAmelCase__ : Any ) -> List[str]: return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase__ ).image_processor def _lowercase ( self : Optional[Any] ) -> Union[str, Any]: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : Dict ) -> List[str]: _a : List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _a : List[Any] = [Image.fromarray(np.moveaxis(UpperCAmelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs def _lowercase ( self : Union[str, Any] ) -> Union[str, Any]: _a : Optional[int] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _a : str = self.get_image_processor(do_normalize=UpperCAmelCase__ , padding_value=1.0 ) _a : Union[str, Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=UpperCAmelCase__ , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , UpperCAmelCase__ ) def _lowercase ( self : List[Any] ) -> str: _a : Union[str, Any] = self.get_image_processor() _a : Dict = SamProcessor(image_processor=UpperCAmelCase__ ) _a : int = self.prepare_image_inputs() _a : List[str] = image_processor(UpperCAmelCase__ , return_tensors="""np""" ) _a : List[str] = processor(images=UpperCAmelCase__ , return_tensors="""np""" ) input_feat_extract.pop("""original_sizes""" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("""reshaped_input_sizes""" ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_tf def _lowercase ( self : Optional[Any] ) -> int: _a : Optional[Any] = self.get_image_processor() _a : Dict = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Any = [tf.ones((1, 3, 5, 5) )] _a : Tuple = [[1764, 2646]] _a : str = [[683, 1024]] _a : Union[str, Any] = processor.post_process_masks(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , return_tensors="""tf""" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) _a : Union[str, Any] = processor.post_process_masks( UpperCAmelCase__ , tf.convert_to_tensor(UpperCAmelCase__ ) , tf.convert_to_tensor(UpperCAmelCase__ ) , return_tensors="""tf""" , ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np _a : List[Any] = [np.ones((1, 3, 5, 5) )] _a : Optional[int] = processor.post_process_masks( UpperCAmelCase__ , np.array(UpperCAmelCase__ ) , np.array(UpperCAmelCase__ ) , return_tensors="""tf""" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) _a : Dict = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): _a : List[Any] = processor.post_process_masks( UpperCAmelCase__ , np.array(UpperCAmelCase__ ) , np.array(UpperCAmelCase__ ) , return_tensors="""tf""" ) @require_vision @require_torchvision class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : str ) -> Optional[Any]: _a : Optional[Any] = tempfile.mkdtemp() _a : Dict = SamImageProcessor() _a : List[str] = SamProcessor(UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) def _lowercase ( self : Any , **UpperCAmelCase__ : Dict ) -> int: return AutoProcessor.from_pretrained(self.tmpdirname , **UpperCAmelCase__ ).image_processor def _lowercase ( self : Tuple ) -> List[Any]: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : str ) -> int: _a : str = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] _a : int = [Image.fromarray(np.moveaxis(UpperCAmelCase__ , 0 , -1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def _lowercase ( self : int ) -> List[Any]: _a : Optional[Any] = self.get_image_processor() _a : Optional[Any] = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Tuple = np.random.randint(0 , 2 , size=(1, 3, 5, 5) ).astype(np.floataa ) _a : str = [tf.convert_to_tensor(UpperCAmelCase__ )] _a : Optional[int] = [torch.tensor(UpperCAmelCase__ )] _a : Union[str, Any] = [[1764, 2646]] _a : List[str] = [[683, 1024]] _a : Optional[int] = processor.post_process_masks( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , return_tensors="""tf""" ) _a : List[str] = processor.post_process_masks( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , return_tensors="""pt""" ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def _lowercase ( self : str ) -> Optional[Any]: _a : List[Any] = self.get_image_processor() _a : Any = SamProcessor(image_processor=UpperCAmelCase__ ) _a : Dict = self.prepare_image_inputs() _a : List[str] = image_processor(UpperCAmelCase__ , return_tensors="""pt""" )["""pixel_values"""].numpy() _a : str = processor(images=UpperCAmelCase__ , return_tensors="""pt""" )["""pixel_values"""].numpy() _a : Optional[Any] = image_processor(UpperCAmelCase__ , return_tensors="""tf""" )["""pixel_values"""].numpy() _a : Optional[int] = processor(images=UpperCAmelCase__ , return_tensors="""tf""" )["""pixel_values"""].numpy() self.assertTrue(np.allclose(UpperCAmelCase__ , UpperCAmelCase__ ) ) self.assertTrue(np.allclose(UpperCAmelCase__ , UpperCAmelCase__ ) ) self.assertTrue(np.allclose(UpperCAmelCase__ , UpperCAmelCase__ ) )
294
1
"""simple docstring""" import unittest import numpy as np def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = None , ): '''simple docstring''' _a : List[Any] = np.shape(UpperCamelCase__ ) _a : Any = np.shape(UpperCamelCase__ ) _a : Union[str, Any] = np.shape(UpperCamelCase__ ) if shape_a[0] != shape_b[0]: _a : int = ( """Expected the same number of rows for A and B. """ F"""Instead found A of size {shape_a} and B of size {shape_b}""" ) raise ValueError(UpperCamelCase__ ) if shape_b[1] != shape_c[1]: _a : Tuple = ( """Expected the same number of columns for B and C. """ F"""Instead found B of size {shape_b} and C of size {shape_c}""" ) raise ValueError(UpperCamelCase__ ) _a : int = pseudo_inv if a_inv is None: try: _a : Optional[int] = np.linalg.inv(UpperCamelCase__ ) except np.linalg.LinAlgError: raise ValueError( """Input matrix A is not invertible. Cannot compute Schur complement.""" ) return mat_c - mat_b.T @ a_inv @ mat_b class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : int ) -> None: _a : str = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) _a : Tuple = np.array([[0, 3], [3, 0], [2, 3]] ) _a : Optional[int] = np.array([[2, 1], [6, 3]] ) _a : Optional[Any] = schur_complement(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) _a : Union[str, Any] = np.block([[a, b], [b.T, c]] ) _a : int = np.linalg.det(UpperCAmelCase__ ) _a : Union[str, Any] = np.linalg.det(UpperCAmelCase__ ) _a : List[Any] = np.linalg.det(UpperCAmelCase__ ) self.assertAlmostEqual(UpperCAmelCase__ , det_a * det_s ) def _lowercase ( self : int ) -> None: _a : Optional[int] = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) _a : Optional[int] = np.array([[0, 3], [3, 0], [2, 3]] ) _a : Union[str, Any] = np.array([[2, 1], [6, 3]] ) with self.assertRaises(UpperCAmelCase__ ): schur_complement(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) def _lowercase ( self : List[Any] ) -> None: _a : Any = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) _a : Dict = np.array([[0, 3], [3, 0], [2, 3]] ) _a : List[Any] = np.array([[2, 1, 3], [6, 3, 5]] ) with self.assertRaises(UpperCAmelCase__ ): schur_complement(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) if __name__ == "__main__": import doctest doctest.testmod() unittest.main()
294
"""simple docstring""" import argparse import gc import json import os import re import torch from huggingface_hub import hf_hub_download from transformers import AutoModelForCausalLM, AutoTokenizer, PreTrainedTokenizerFast, RwkvConfig from transformers.modeling_utils import WEIGHTS_INDEX_NAME, shard_checkpoint _snake_case = { '169M': 12, '430M': 24, '1B5': 24, '3B': 32, '7B': 32, '14B': 40, } _snake_case = { '169M': 768, '430M': 1024, '1B5': 2048, '3B': 2560, '7B': 4096, '14B': 5120, } def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : int = list(state_dict.keys() ) for name in state_dict_keys: _a : str = state_dict.pop(UpperCamelCase__ ) # emb -> embedding if name.startswith("""emb.""" ): _a : Dict = name.replace("""emb.""" , """embeddings.""" ) # ln_0 -> pre_ln (only present at block 0) if name.startswith("""blocks.0.ln0""" ): _a : Dict = name.replace("""blocks.0.ln0""" , """blocks.0.pre_ln""" ) # att -> attention _a : Any = re.sub(R"""blocks\.(\d+)\.att""" , R"""blocks.\1.attention""" , UpperCamelCase__ ) # ffn -> feed_forward _a : int = re.sub(R"""blocks\.(\d+)\.ffn""" , R"""blocks.\1.feed_forward""" , UpperCamelCase__ ) # time_mix_k -> time_mix_key and reshape if name.endswith(""".time_mix_k""" ): _a : List[str] = name.replace(""".time_mix_k""" , """.time_mix_key""" ) # time_mix_v -> time_mix_value and reshape if name.endswith(""".time_mix_v""" ): _a : Tuple = name.replace(""".time_mix_v""" , """.time_mix_value""" ) # time_mix_r -> time_mix_key and reshape if name.endswith(""".time_mix_r""" ): _a : Dict = name.replace(""".time_mix_r""" , """.time_mix_receptance""" ) if name != "head.weight": _a : Optional[int] = """rwkv.""" + name _a : Any = weight return state_dict def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=False , UpperCamelCase__=None ): '''simple docstring''' # 1. If possible, build the tokenizer. if tokenizer_file is None: print("""No `--tokenizer_file` provided, we will use the default tokenizer.""" ) _a : Tuple = 5_0_2_7_7 _a : str = AutoTokenizer.from_pretrained("""EleutherAI/gpt-neox-20b""" ) else: _a : int = PreTrainedTokenizerFast(tokenizer_file=UpperCamelCase__ ) _a : int = len(UpperCamelCase__ ) tokenizer.save_pretrained(UpperCamelCase__ ) # 2. Build the config _a : Optional[Any] = list(NUM_HIDDEN_LAYERS_MAPPING.keys() ) if size is None: # Try to infer size from the checkpoint name for candidate in possible_sizes: if candidate in checkpoint_file: _a : Tuple = candidate break if size is None: raise ValueError("""Could not infer the size, please provide it with the `--size` argument.""" ) if size not in possible_sizes: raise ValueError(F"""`size` should be one of {possible_sizes}, got {size}.""" ) _a : List[Any] = RwkvConfig( vocab_size=UpperCamelCase__ , num_hidden_layers=NUM_HIDDEN_LAYERS_MAPPING[size] , hidden_size=HIDEN_SIZE_MAPPING[size] , ) config.save_pretrained(UpperCamelCase__ ) # 3. Download model file then convert state_dict _a : str = hf_hub_download(UpperCamelCase__ , UpperCamelCase__ ) _a : int = torch.load(UpperCamelCase__ , map_location="""cpu""" ) _a : List[str] = convert_state_dict(UpperCamelCase__ ) # 4. Split in shards and save _a , _a : List[str] = shard_checkpoint(UpperCamelCase__ ) for shard_file, shard in shards.items(): torch.save(UpperCamelCase__ , os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) if index is not None: _a : Optional[Any] = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) # Save the index as well with open(UpperCamelCase__ , """w""" , encoding="""utf-8""" ) as f: _a : Dict = json.dumps(UpperCamelCase__ , indent=2 , sort_keys=UpperCamelCase__ ) + """\n""" f.write(UpperCamelCase__ ) # 5. Clean up shards (for some reason the file PyTorch saves take the same space as the whole state_dict print( """Cleaning up shards. This may error with an OOM error, it this is the case don't worry you still have converted the model.""" ) _a : List[Any] = list(shards.keys() ) del state_dict del shards gc.collect() for shard_file in shard_files: _a : Any = torch.load(os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) torch.save({k: v.cpu().clone() for k, v in state_dict.items()} , os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) del state_dict gc.collect() if push_to_hub: if model_name is None: raise ValueError("""Please provide a `model_name` to push the model to the Hub.""" ) _a : Dict = AutoModelForCausalLM.from_pretrained(UpperCamelCase__ ) model.push_to_hub(UpperCamelCase__ , max_shard_size="""2GB""" ) tokenizer.push_to_hub(UpperCamelCase__ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '--repo_id', default=None, type=str, required=True, help='Repo ID from which to pull the checkpoint.' ) parser.add_argument( '--checkpoint_file', default=None, type=str, required=True, help='Name of the checkpoint file in the repo.' ) parser.add_argument( '--output_dir', default=None, type=str, required=True, help='Where to save the converted model.' ) parser.add_argument( '--tokenizer_file', default=None, type=str, help='Path to the tokenizer file to use (if not provided, only the model is converted).', ) parser.add_argument( '--size', default=None, type=str, help='Size of the model. Will be inferred from the `checkpoint_file` if not passed.', ) parser.add_argument( '--push_to_hub', action='store_true', help='Push to the Hub the converted model.', ) parser.add_argument( '--model_name', default=None, type=str, help='Name of the pushed model on the Hub, including the username / organization.', ) _snake_case = parser.parse_args() convert_rmkv_checkpoint_to_hf_format( args.repo_id, args.checkpoint_file, args.output_dir, size=args.size, tokenizer_file=args.tokenizer_file, push_to_hub=args.push_to_hub, model_name=args.model_name, )
294
1
"""simple docstring""" import unittest from transformers.utils.backbone_utils import ( BackboneMixin, get_aligned_output_features_output_indices, verify_out_features_out_indices, ) class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : Optional[Any] ) -> int: _a : Optional[int] = ["""a""", """b""", """c"""] # Defaults to last layer if both are None _a , _a : Dict = get_aligned_output_features_output_indices(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) self.assertEqual(UpperCAmelCase__ , ["""c"""] ) self.assertEqual(UpperCAmelCase__ , [2] ) # Out indices set to match out features _a , _a : Optional[Any] = get_aligned_output_features_output_indices(["""a""", """c"""] , UpperCAmelCase__ , UpperCAmelCase__ ) self.assertEqual(UpperCAmelCase__ , ["""a""", """c"""] ) self.assertEqual(UpperCAmelCase__ , [0, 2] ) # Out features set to match out indices _a , _a : List[Any] = get_aligned_output_features_output_indices(UpperCAmelCase__ , [0, 2] , UpperCAmelCase__ ) self.assertEqual(UpperCAmelCase__ , ["""a""", """c"""] ) self.assertEqual(UpperCAmelCase__ , [0, 2] ) # Out features selected from negative indices _a , _a : str = get_aligned_output_features_output_indices(UpperCAmelCase__ , [-3, -1] , UpperCAmelCase__ ) self.assertEqual(UpperCAmelCase__ , ["""a""", """c"""] ) self.assertEqual(UpperCAmelCase__ , [-3, -1] ) def _lowercase ( self : Tuple ) -> int: # Stage names must be set with self.assertRaises(UpperCAmelCase__ ): verify_out_features_out_indices(["""a""", """b"""] , (0, 1) , UpperCAmelCase__ ) # Out features must be a list with self.assertRaises(UpperCAmelCase__ ): verify_out_features_out_indices(("""a""", """b""") , (0, 1) , ["""a""", """b"""] ) # Out features must be a subset of stage names with self.assertRaises(UpperCAmelCase__ ): verify_out_features_out_indices(["""a""", """b"""] , (0, 1) , ["""a"""] ) # Out indices must be a list or tuple with self.assertRaises(UpperCAmelCase__ ): verify_out_features_out_indices(UpperCAmelCase__ , 0 , ["""a""", """b"""] ) # Out indices must be a subset of stage names with self.assertRaises(UpperCAmelCase__ ): verify_out_features_out_indices(UpperCAmelCase__ , (0, 1) , ["""a"""] ) # Out features and out indices must be the same length with self.assertRaises(UpperCAmelCase__ ): verify_out_features_out_indices(["""a""", """b"""] , (0,) , ["""a""", """b""", """c"""] ) # Out features should match out indices with self.assertRaises(UpperCAmelCase__ ): verify_out_features_out_indices(["""a""", """b"""] , (0, 2) , ["""a""", """b""", """c"""] ) # Out features and out indices should be in order with self.assertRaises(UpperCAmelCase__ ): verify_out_features_out_indices(["""b""", """a"""] , (0, 1) , ["""a""", """b"""] ) # Check passes with valid inputs verify_out_features_out_indices(["""a""", """b""", """d"""] , (0, 1, -1) , ["""a""", """b""", """c""", """d"""] ) def _lowercase ( self : Any ) -> Tuple: _a : Optional[Any] = BackboneMixin() _a : Dict = ["""a""", """b""", """c"""] _a : Any = ["""a""", """c"""] _a : Union[str, Any] = [0, 2] # Check that the output features and indices are set correctly self.assertEqual(backbone.out_features , ["""a""", """c"""] ) self.assertEqual(backbone.out_indices , [0, 2] ) # Check out features and indices are updated correctly _a : List[Any] = ["""a""", """b"""] self.assertEqual(backbone.out_features , ["""a""", """b"""] ) self.assertEqual(backbone.out_indices , [0, 1] ) _a : Union[str, Any] = [-3, -1] self.assertEqual(backbone.out_features , ["""a""", """c"""] ) self.assertEqual(backbone.out_indices , [-3, -1] )
294
"""simple docstring""" import shutil import tempfile import unittest from transformers import ClapFeatureExtractor, ClapProcessor, RobertaTokenizer, RobertaTokenizerFast from transformers.testing_utils import require_sentencepiece, require_torchaudio from .test_feature_extraction_clap import floats_list @require_torchaudio @require_sentencepiece class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : int ) -> List[str]: _a : Any = """laion/clap-htsat-unfused""" _a : Union[str, Any] = tempfile.mkdtemp() def _lowercase ( self : List[Any] , **UpperCAmelCase__ : Any ) -> Dict: return RobertaTokenizer.from_pretrained(self.checkpoint , **UpperCAmelCase__ ) def _lowercase ( self : List[Any] , **UpperCAmelCase__ : List[str] ) -> int: return ClapFeatureExtractor.from_pretrained(self.checkpoint , **UpperCAmelCase__ ) def _lowercase ( self : List[Any] ) -> Tuple: shutil.rmtree(self.tmpdirname ) def _lowercase ( self : List[str] ) -> Optional[int]: _a : List[str] = self.get_tokenizer() _a : Any = self.get_feature_extractor() _a : Optional[Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) _a : List[str] = ClapProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCAmelCase__ ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , UpperCAmelCase__ ) def _lowercase ( self : Dict ) -> Optional[int]: _a : Tuple = ClapProcessor(tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() ) processor.save_pretrained(self.tmpdirname ) _a : Dict = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) _a : Union[str, Any] = self.get_feature_extractor(do_normalize=UpperCAmelCase__ , padding_value=1.0 ) _a : Union[str, Any] = ClapProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=UpperCAmelCase__ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , UpperCAmelCase__ ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.feature_extractor , UpperCAmelCase__ ) def _lowercase ( self : List[str] ) -> Optional[Any]: _a : Optional[int] = self.get_feature_extractor() _a : Tuple = self.get_tokenizer() _a : List[Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) _a : Any = floats_list((3, 1000) ) _a : List[Any] = feature_extractor(UpperCAmelCase__ , return_tensors="""np""" ) _a : List[str] = processor(audios=UpperCAmelCase__ , return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def _lowercase ( self : Tuple ) -> Optional[int]: _a : List[str] = self.get_feature_extractor() _a : Any = self.get_tokenizer() _a : Any = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) _a : Optional[int] = """This is a test string""" _a : Tuple = processor(text=UpperCAmelCase__ ) _a : int = tokenizer(UpperCAmelCase__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _lowercase ( self : List[Any] ) -> Any: _a : str = self.get_feature_extractor() _a : List[str] = self.get_tokenizer() _a : List[Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) _a : Any = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _a : Dict = processor.batch_decode(UpperCAmelCase__ ) _a : Any = tokenizer.batch_decode(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def _lowercase ( self : Dict ) -> List[str]: _a : str = self.get_feature_extractor() _a : Optional[Any] = self.get_tokenizer() _a : Union[str, Any] = ClapProcessor(tokenizer=UpperCAmelCase__ , feature_extractor=UpperCAmelCase__ ) self.assertListEqual( processor.model_input_names[2:] , feature_extractor.model_input_names , msg="""`processor` and `feature_extractor` model input names do not match""" , )
294
1
"""simple docstring""" from argparse import ArgumentParser from . import BaseTransformersCLICommand def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class UpperCamelCase ( snake_case_ ): @staticmethod def _lowercase ( UpperCAmelCase__ : ArgumentParser ) -> Optional[Any]: _a : Optional[Any] = parser.add_parser("""download""" ) download_parser.add_argument( """--cache-dir""" , type=UpperCAmelCase__ , default=UpperCAmelCase__ , help="""Path to location to store the models""" ) download_parser.add_argument( """--force""" , action="""store_true""" , help="""Force the model to be download even if already in cache-dir""" ) download_parser.add_argument( """--trust-remote-code""" , action="""store_true""" , help="""Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you've reviewed the code as it will execute on your local machine""" , ) download_parser.add_argument("""model""" , type=UpperCAmelCase__ , help="""Name of the model to download""" ) download_parser.set_defaults(func=UpperCAmelCase__ ) def __init__( self : Optional[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : str , UpperCAmelCase__ : bool , UpperCAmelCase__ : bool ) -> int: _a : Dict = model _a : Dict = cache _a : Optional[Any] = force _a : List[str] = trust_remote_code def _lowercase ( self : List[Any] ) -> List[str]: from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code )
294
"""simple docstring""" import warnings from ...utils import logging from .image_processing_clip import CLIPImageProcessor _snake_case = logging.get_logger(__name__) class UpperCamelCase ( snake_case_ ): def __init__( self : Any , *UpperCAmelCase__ : Optional[Any] , **UpperCAmelCase__ : Tuple ) -> None: warnings.warn( """The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use CLIPImageProcessor instead.""" , UpperCAmelCase__ , ) super().__init__(*UpperCAmelCase__ , **UpperCAmelCase__ )
294
1
"""simple docstring""" import json import os from typing import Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = {'vocab_file': 'vocab.json'} _snake_case = { 'vocab_file': { 'mgp-str': 'https://huggingface.co/alibaba-damo/mgp-str-base/blob/main/vocab.json', } } _snake_case = {'mgp-str': 27} class UpperCamelCase ( snake_case_ ): UpperCamelCase : List[str] = VOCAB_FILES_NAMES UpperCamelCase : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : Optional[int] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : List[Any]="[GO]" , UpperCAmelCase__ : Tuple="[GO]" , UpperCAmelCase__ : Optional[int]="[s]" , UpperCAmelCase__ : int="[GO]" , **UpperCAmelCase__ : Dict ) -> int: super().__init__( unk_token=UpperCAmelCase__ , bos_token=UpperCAmelCase__ , eos_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , **UpperCAmelCase__ , ) with open(UpperCAmelCase__ , encoding="""utf-8""" ) as vocab_handle: _a : int = json.load(UpperCAmelCase__ ) _a : Optional[int] = {v: k for k, v in self.vocab.items()} @property def _lowercase ( self : Dict ) -> Union[str, Any]: return len(self.vocab ) def _lowercase ( self : Union[str, Any] ) -> str: return dict(self.vocab , **self.added_tokens_encoder ) def _lowercase ( self : Dict , UpperCAmelCase__ : str ) -> Union[str, Any]: _a : Tuple = [] for s in text: char_tokens.extend(UpperCAmelCase__ ) return char_tokens def _lowercase ( self : List[Any] , UpperCAmelCase__ : str ) -> Dict: return self.vocab.get(UpperCAmelCase__ , self.vocab.get(self.unk_token ) ) def _lowercase ( self : Optional[int] , UpperCAmelCase__ : Tuple ) -> List[Any]: return self.decoder.get(UpperCAmelCase__ ) def _lowercase ( self : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(UpperCAmelCase__ ): logger.error("""Vocabulary path ({}) should be a directory""".format(UpperCAmelCase__ ) ) return _a : Tuple = os.path.join( UpperCAmelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) with open(UpperCAmelCase__ , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.vocab , indent=2 , sort_keys=UpperCAmelCase__ , ensure_ascii=UpperCAmelCase__ ) + """\n""" ) return (vocab_file,)
294
"""simple docstring""" import unittest import numpy as np def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = None , ): '''simple docstring''' _a : List[Any] = np.shape(UpperCamelCase__ ) _a : Any = np.shape(UpperCamelCase__ ) _a : Union[str, Any] = np.shape(UpperCamelCase__ ) if shape_a[0] != shape_b[0]: _a : int = ( """Expected the same number of rows for A and B. """ F"""Instead found A of size {shape_a} and B of size {shape_b}""" ) raise ValueError(UpperCamelCase__ ) if shape_b[1] != shape_c[1]: _a : Tuple = ( """Expected the same number of columns for B and C. """ F"""Instead found B of size {shape_b} and C of size {shape_c}""" ) raise ValueError(UpperCamelCase__ ) _a : int = pseudo_inv if a_inv is None: try: _a : Optional[int] = np.linalg.inv(UpperCamelCase__ ) except np.linalg.LinAlgError: raise ValueError( """Input matrix A is not invertible. Cannot compute Schur complement.""" ) return mat_c - mat_b.T @ a_inv @ mat_b class UpperCamelCase ( unittest.TestCase ): def _lowercase ( self : int ) -> None: _a : str = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) _a : Tuple = np.array([[0, 3], [3, 0], [2, 3]] ) _a : Optional[int] = np.array([[2, 1], [6, 3]] ) _a : Optional[Any] = schur_complement(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) _a : Union[str, Any] = np.block([[a, b], [b.T, c]] ) _a : int = np.linalg.det(UpperCAmelCase__ ) _a : Union[str, Any] = np.linalg.det(UpperCAmelCase__ ) _a : List[Any] = np.linalg.det(UpperCAmelCase__ ) self.assertAlmostEqual(UpperCAmelCase__ , det_a * det_s ) def _lowercase ( self : int ) -> None: _a : Optional[int] = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) _a : Optional[int] = np.array([[0, 3], [3, 0], [2, 3]] ) _a : Union[str, Any] = np.array([[2, 1], [6, 3]] ) with self.assertRaises(UpperCAmelCase__ ): schur_complement(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) def _lowercase ( self : List[Any] ) -> None: _a : Any = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) _a : Dict = np.array([[0, 3], [3, 0], [2, 3]] ) _a : List[Any] = np.array([[2, 1, 3], [6, 3, 5]] ) with self.assertRaises(UpperCAmelCase__ ): schur_complement(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) if __name__ == "__main__": import doctest doctest.testmod() unittest.main()
294
1
"""simple docstring""" import math def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' return math.pow(UpperCamelCase__ , 2 ) - a def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' return 2 * x def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : List[str] = 2.0 while start <= a: _a : Dict = math.pow(UpperCamelCase__ , 2 ) return start def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ = 9_9_9_9 , UpperCamelCase__ = 0.00_000_000_000_001 ): '''simple docstring''' if a < 0: raise ValueError("""math domain error""" ) _a : List[str] = get_initial_point(UpperCamelCase__ ) for _ in range(UpperCamelCase__ ): _a : List[Any] = value _a : str = value - fx(UpperCamelCase__ , UpperCamelCase__ ) / fx_derivative(UpperCamelCase__ ) if abs(prev_value - value ) < tolerance: return value return value if __name__ == "__main__": from doctest import testmod testmod()
294
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _snake_case = logging.get_logger(__name__) _snake_case = torch.device('cpu') def lowerCAmelCase__ ( ): '''simple docstring''' _a : int = """http://images.cocodataset.org/val2017/000000039769.jpg""" _a : Dict = Image.open(requests.get(UpperCamelCase__ , stream=UpperCamelCase__ ).raw ) return im def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' if swiftformer_name == "swiftformer_xs": return torch.tensor([-2.17_03e00, 2.11_07e00, -2.08_11e00, 8.86_85e-01, 2.43_60e-01] ) elif swiftformer_name == "swiftformer_s": return torch.tensor([3.96_36e-01, 2.34_78e-01, -1.69_63e00, -1.73_81e00, -8.63_37e-01] ) elif swiftformer_name == "swiftformer_l1": return torch.tensor([-4.27_68e-01, -4.74_29e-01, -1.08_97e00, -1.02_48e00, 3.55_23e-02] ) elif swiftformer_name == "swiftformer_l3": return torch.tensor([-2.53_30e-01, 2.42_11e-01, -6.01_85e-01, -8.27_89e-01, -6.04_46e-02] ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Any = dct.pop(UpperCamelCase__ ) _a : Dict = val def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : Tuple = [] for k in state_dict.keys(): _a : Any = k if ".pwconv" in k: _a : int = k_new.replace(""".pwconv""" , """.point_wise_conv""" ) if ".dwconv" in k: _a : List[str] = k_new.replace(""".dwconv""" , """.depth_wise_conv""" ) if ".Proj." in k: _a : Optional[int] = k_new.replace(""".Proj.""" , """.proj.""" ) if "patch_embed" in k_new: _a : Tuple = k_new.replace("""patch_embed""" , """swiftformer.patch_embed.patch_embedding""" ) if "network" in k_new: _a : int = k_new.split(""".""" ) if ls[2].isdigit(): _a : Union[str, Any] = """swiftformer.encoder.network.""" + ls[1] + """.blocks.""" + ls[2] + """.""" + """.""".join(ls[3:] ) else: _a : Tuple = k_new.replace("""network""" , """swiftformer.encoder.network""" ) rename_keys.append((k, k_new) ) return rename_keys @torch.no_grad() def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Tuple = SwiftFormerConfig() # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size _a : Optional[int] = 1_0_0_0 _a : Optional[Any] = """huggingface/label-files""" _a : Optional[Any] = """imagenet-1k-id2label.json""" _a : List[str] = json.load(open(hf_hub_download(UpperCamelCase__ , UpperCamelCase__ , repo_type="""dataset""" ) , """r""" ) ) _a : Optional[Any] = {int(UpperCamelCase__ ): v for k, v in idalabel.items()} _a : Dict = idalabel _a : Optional[int] = {v: k for k, v in idalabel.items()} # size of the architecture if swiftformer_name == "swiftformer_xs": _a : Any = [3, 3, 6, 4] _a : int = [4_8, 5_6, 1_1_2, 2_2_0] elif swiftformer_name == "swiftformer_s": _a : Any = [3, 3, 9, 6] _a : List[str] = [4_8, 6_4, 1_6_8, 2_2_4] elif swiftformer_name == "swiftformer_l1": _a : List[Any] = [4, 3, 1_0, 5] _a : Optional[int] = [4_8, 9_6, 1_9_2, 3_8_4] elif swiftformer_name == "swiftformer_l3": _a : List[Any] = [4, 4, 1_2, 6] _a : Optional[Any] = [6_4, 1_2_8, 3_2_0, 5_1_2] # load state_dict of original model, remove and rename some keys if original_ckpt: if original_ckpt.startswith("""https""" ): _a : Tuple = torch.hub.load_state_dict_from_url(UpperCamelCase__ , map_location="""cpu""" , check_hash=UpperCamelCase__ ) else: _a : Dict = torch.load(UpperCamelCase__ , map_location="""cpu""" ) _a : int = checkpoint _a : Optional[Any] = create_rename_keys(UpperCamelCase__ ) for rename_key_src, rename_key_dest in rename_keys: rename_key(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # load HuggingFace model _a : Any = SwiftFormerForImageClassification(UpperCamelCase__ ).eval() hf_model.load_state_dict(UpperCamelCase__ ) # prepare test inputs _a : Any = prepare_img() _a : Union[str, Any] = ViTImageProcessor.from_pretrained("""preprocessor_config""" ) _a : Optional[int] = processor(images=UpperCamelCase__ , return_tensors="""pt""" ) # compare outputs from both models _a : Dict = get_expected_output(UpperCamelCase__ ) _a : int = hf_model(inputs["""pixel_values"""] ).logits assert hf_logits.shape == torch.Size([1, 1_0_0_0] ) assert torch.allclose(hf_logits[0, 0:5] , UpperCamelCase__ , atol=1e-3 ) Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) print(F"""Saving model {swiftformer_name} to {pytorch_dump_folder_path}""" ) hf_model.save_pretrained(UpperCamelCase__ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '--swiftformer_name', default='swiftformer_xs', choices=['swiftformer_xs', 'swiftformer_s', 'swiftformer_l1', 'swiftformer_l3'], type=str, help='Name of the SwiftFormer model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default='./converted_outputs/', type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument('--original_ckpt', default=None, type=str, help='Path to the original model checkpoint.') _snake_case = parser.parse_args() convert_swiftformer_checkpoint(args.swiftformer_name, args.pytorch_dump_folder_path, args.original_ckpt)
294
1
"""simple docstring""" import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Value from .base import TaskTemplate @dataclass(frozen=snake_case_ ) class UpperCamelCase ( snake_case_ ): # `task` is not a ClassVar since we want it to be part of the `asdict` output for JSON serialization UpperCamelCase : str = field(default='''text-classification''' , metadata={'''include_in_asdict_even_if_is_default''': True} ) UpperCamelCase : ClassVar[Features] = Features({'''text''': Value('''string''' )} ) UpperCamelCase : ClassVar[Features] = Features({'''labels''': ClassLabel} ) UpperCamelCase : str = "text" UpperCamelCase : str = "labels" def _lowercase ( self : Union[str, Any] , UpperCAmelCase__ : Dict ) -> Optional[int]: if self.label_column not in features: raise ValueError(f"""Column {self.label_column} is not present in features.""" ) if not isinstance(features[self.label_column] , UpperCAmelCase__ ): raise ValueError(f"""Column {self.label_column} is not a ClassLabel.""" ) _a : Optional[int] = copy.deepcopy(self ) _a : Tuple = self.label_schema.copy() _a : Any = features[self.label_column] _a : Tuple = label_schema return task_template @property def _lowercase ( self : List[str] ) -> Dict[str, str]: return { self.text_column: "text", self.label_column: "labels", }
294
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case = { 'configuration_perceiver': ['PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PerceiverConfig', 'PerceiverOnnxConfig'], 'tokenization_perceiver': ['PerceiverTokenizer'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ['PerceiverFeatureExtractor'] _snake_case = ['PerceiverImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST', 'PerceiverForImageClassificationConvProcessing', 'PerceiverForImageClassificationFourier', 'PerceiverForImageClassificationLearned', 'PerceiverForMaskedLM', 'PerceiverForMultimodalAutoencoding', 'PerceiverForOpticalFlow', 'PerceiverForSequenceClassification', 'PerceiverLayer', 'PerceiverModel', 'PerceiverPreTrainedModel', ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
294
1
"""simple docstring""" from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def lowerCAmelCase__ ( ): '''simple docstring''' _a : Any = HfArgumentParser(UpperCamelCase__ ) _a : int = parser.parse_args_into_dataclasses()[0] _a : Tuple = TensorFlowBenchmark(args=UpperCamelCase__ ) try: _a : Dict = parser.parse_args_into_dataclasses()[0] except ValueError as e: _a : List[Any] = """Arg --no_{0} is no longer used, please use --no-{0} instead.""" _a : List[str] = """ """.join(str(UpperCamelCase__ ).split(""" """ )[:-1] ) _a : Union[str, Any] = """""" _a : Dict = eval(str(UpperCamelCase__ ).split(""" """ )[-1] ) _a : Optional[Any] = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:] ) else: wrong_args.append(UpperCamelCase__ ) if len(UpperCamelCase__ ) > 0: _a : Dict = full_error_msg + begin_error_msg + str(UpperCamelCase__ ) raise ValueError(UpperCamelCase__ ) benchmark.run() if __name__ == "__main__": main()
294
"""simple docstring""" import cva import numpy as np class UpperCamelCase : def __init__( self : Optional[int] , UpperCAmelCase__ : float , UpperCAmelCase__ : int ) -> Dict: if k in (0.0_4, 0.0_6): _a : List[str] = k _a : List[Any] = window_size else: raise ValueError("""invalid k value""" ) def __str__( self : Dict ) -> str: return str(self.k ) def _lowercase ( self : int , UpperCAmelCase__ : str ) -> tuple[cva.Mat, list[list[int]]]: _a : Dict = cva.imread(UpperCAmelCase__ , 0 ) _a , _a : List[Any] = img.shape _a : list[list[int]] = [] _a : List[Any] = img.copy() _a : int = cva.cvtColor(UpperCAmelCase__ , cva.COLOR_GRAY2RGB ) _a , _a : Any = np.gradient(UpperCAmelCase__ ) _a : Tuple = dx**2 _a : Union[str, Any] = dy**2 _a : Union[str, Any] = dx * dy _a : int = 0.0_4 _a : List[str] = self.window_size // 2 for y in range(UpperCAmelCase__ , h - offset ): for x in range(UpperCAmelCase__ , w - offset ): _a : str = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _a : List[Any] = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _a : Tuple = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _a : Any = (wxx * wyy) - (wxy**2) _a : Tuple = wxx + wyy _a : Any = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 255 ) return color_img, corner_list if __name__ == "__main__": _snake_case = HarrisCorner(0.04, 3) _snake_case , _snake_case = edge_detect.detect('path_to_image') cva.imwrite('detect.png', color_img)
294
1
"""simple docstring""" import argparse import requests import torch # pip3 install salesforce-lavis # I'm actually installing a slightly modified version: pip3 install git+https://github.com/nielsrogge/LAVIS.git@fix_lavis_float32 (there's also the fix_lavis branch) # also note: to convert Vicuna checkpoints, we had to include /home/niels/python_projects/checkpoints/FastChat/vicuna-7b in lavis/configs/models/blip2/blip2_instruct_vicuna7b.yaml # same for Vicuna-13b from lavis.models import load_model_and_preprocess from PIL import Image from transformers import ( AutoTokenizer, BlipImageProcessor, InstructBlipConfig, InstructBlipForConditionalGeneration, InstructBlipProcessor, InstructBlipQFormerConfig, InstructBlipVisionConfig, LlamaConfig, LlamaTokenizerFast, TaConfig, TaTokenizerFast, ) from transformers.utils.constants import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD def lowerCAmelCase__ ( ): '''simple docstring''' _a : List[Any] = """https://raw.githubusercontent.com/salesforce/LAVIS/main/docs/_static/Confusing-Pictures.jpg""" _a : List[str] = Image.open(requests.get(UpperCamelCase__ , stream=UpperCamelCase__ ).raw ).convert("""RGB""" ) return image def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : Tuple = [] # fmt: off # vision encoder rename_keys.append(("""visual_encoder.cls_token""", """vision_model.embeddings.class_embedding""") ) rename_keys.append(("""visual_encoder.pos_embed""", """vision_model.embeddings.position_embedding""") ) rename_keys.append(("""visual_encoder.patch_embed.proj.weight""", """vision_model.embeddings.patch_embedding.weight""") ) rename_keys.append(("""visual_encoder.patch_embed.proj.bias""", """vision_model.embeddings.patch_embedding.bias""") ) rename_keys.append(("""ln_vision.weight""", """vision_model.post_layernorm.weight""") ) rename_keys.append(("""ln_vision.bias""", """vision_model.post_layernorm.bias""") ) for i in range(config.vision_config.num_hidden_layers ): rename_keys.append((F"""visual_encoder.blocks.{i}.norm1.weight""", F"""vision_model.encoder.layers.{i}.layer_norm1.weight""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.norm1.bias""", F"""vision_model.encoder.layers.{i}.layer_norm1.bias""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.norm2.weight""", F"""vision_model.encoder.layers.{i}.layer_norm2.weight""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.norm2.bias""", F"""vision_model.encoder.layers.{i}.layer_norm2.bias""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.attn.qkv.weight""", F"""vision_model.encoder.layers.{i}.self_attn.qkv.weight""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.attn.proj.weight""", F"""vision_model.encoder.layers.{i}.self_attn.projection.weight""",) ) rename_keys.append((F"""visual_encoder.blocks.{i}.attn.proj.bias""", F"""vision_model.encoder.layers.{i}.self_attn.projection.bias""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.mlp.fc1.weight""", F"""vision_model.encoder.layers.{i}.mlp.fc1.weight""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.mlp.fc1.bias""", F"""vision_model.encoder.layers.{i}.mlp.fc1.bias""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.mlp.fc2.weight""", F"""vision_model.encoder.layers.{i}.mlp.fc2.weight""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.mlp.fc2.bias""", F"""vision_model.encoder.layers.{i}.mlp.fc2.bias""") ) # QFormer rename_keys.append(("""Qformer.bert.embeddings.LayerNorm.weight""", """qformer.embeddings.layernorm.weight""") ) rename_keys.append(("""Qformer.bert.embeddings.LayerNorm.bias""", """qformer.embeddings.layernorm.bias""") ) # fmt: on return rename_keys def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Dict = dct.pop(UpperCamelCase__ ) _a : Union[str, Any] = val def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' for i in range(config.vision_config.num_hidden_layers ): # read in original q and v biases _a : int = state_dict.pop(F"""visual_encoder.blocks.{i}.attn.q_bias""" ) _a : int = state_dict.pop(F"""visual_encoder.blocks.{i}.attn.v_bias""" ) # next, set bias in the state dict _a : Optional[int] = torch.cat((q_bias, torch.zeros_like(UpperCamelCase__ , requires_grad=UpperCamelCase__ ), v_bias) ) _a : Optional[int] = qkv_bias def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : Tuple = 3_6_4 if """coco""" in model_name else 2_2_4 _a : int = InstructBlipVisionConfig(image_size=UpperCamelCase__ ).to_dict() # make sure the models have proper bos_token_id and eos_token_id set (important for generation) # seems like flan-T5 models don't have bos_token_id properly set? if "t5-xl" in model_name: _a : Optional[int] = TaConfig.from_pretrained("""google/flan-t5-xl""" , dense_act_fn="""gelu""" , bos_token_id=1 ).to_dict() elif "t5-xxl" in model_name: _a : List[str] = TaConfig.from_pretrained("""google/flan-t5-xxl""" , dense_act_fn="""gelu""" , bos_token_id=1 ).to_dict() elif "vicuna-7b" in model_name: _a : int = LlamaConfig.from_pretrained("""decapoda-research/llama-7b-hf""" , vocab_size=3_2_0_0_1 ).to_dict() elif "vicuna-13b" in model_name: _a : Any = LlamaConfig.from_pretrained("""decapoda-research/llama-13b-hf""" , vocab_size=3_2_0_0_1 ).to_dict() else: raise ValueError("""Model name not supported""" ) # the authors add one special "[DEC]" token to the vocab of Q-Former, hence vocab size = 30522 + 1 _a : List[str] = InstructBlipQFormerConfig(vocab_size=3_0_5_2_3 ).to_dict() _a : List[Any] = InstructBlipConfig(vision_config=UpperCamelCase__ , text_config=UpperCamelCase__ , qformer_config=UpperCamelCase__ ) return config, image_size @torch.no_grad() def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__=None , UpperCamelCase__=False ): '''simple docstring''' _a : Optional[Any] = AutoTokenizer.from_pretrained("""bert-base-uncased""" , truncation_side="""left""" ) qformer_tokenizer.add_special_tokens({"""bos_token""": """[DEC]"""} ) if "t5" in model_name: _a : Union[str, Any] = TaTokenizerFast.from_pretrained("""google/flan-t5-xl""" , truncation_side="""left""" ) elif "vicuna" in model_name: # the following was used in the original implementation: # tokenizer = LlamaTokenizer.from_pretrained("huggyllama/llama-7b", use_fast=False, truncation_side="left") # tokenizer.add_special_tokens({"pad_token": "[PAD]"}) # tokenizer.add_special_tokens({"bos_token": "</s>"}) # tokenizer.add_special_tokens({"eos_token": "</s>"}) # tokenizer.add_special_tokens({"unk_token": "</s>"}) _a : List[str] = LlamaTokenizerFast.from_pretrained( """huggyllama/llama-7b""" , truncation_side="""left""" , bos_token="""</s>""" , unk_token="""</s>""" ) tokenizer.add_special_tokens({"""pad_token""": """[PAD]"""} ) _a , _a : Tuple = get_blipa_config(UpperCamelCase__ ) _a : Optional[Any] = InstructBlipForConditionalGeneration(UpperCamelCase__ ).eval() _a : List[Any] = { """instructblip-vicuna-7b""": ("""blip2_vicuna_instruct""", """vicuna7b"""), """instructblip-vicuna-13b""": ("""blip2_vicuna_instruct""", """vicuna13b"""), """instructblip-flan-t5-xl""": ("""blip2_t5_instruct""", """flant5xl"""), """instructblip-flan-t5-xxl""": ("""blip2_t5_instruct""", """flant5xxl"""), } _a , _a : Union[str, Any] = model_name_to_original[model_name] # load original model print("""Loading original model...""" ) _a : Union[str, Any] = """cuda:1""" if torch.cuda.is_available() else """cpu""" _a : Any = """cuda:2""" if torch.cuda.is_available() else """cpu""" _a , _a , _a : List[str] = load_model_and_preprocess( name=UpperCamelCase__ , model_type=UpperCamelCase__ , is_eval=UpperCamelCase__ , device=UpperCamelCase__ ) original_model.eval() print("""Done!""" ) # update state dict keys _a : Tuple = original_model.state_dict() _a : Dict = create_rename_keys(UpperCamelCase__ ) for src, dest in rename_keys: rename_key(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # some keys can be renamed efficiently for key, val in state_dict.copy().items(): _a : Tuple = state_dict.pop(UpperCamelCase__ ) if key.startswith("""Qformer.bert""" ): _a : Any = key.replace("""Qformer.bert""" , """qformer""" ) if "attention.self" in key: _a : Tuple = key.replace("""self""" , """attention""" ) if "llm_proj" in key: _a : Optional[int] = key.replace("""llm_proj""" , """language_projection""" ) if "t5_proj" in key: _a : Optional[Any] = key.replace("""t5_proj""" , """language_projection""" ) if key.startswith("""llm_model""" ): _a : List[Any] = key.replace("""llm_model""" , """language_model""" ) if key.startswith("""t5""" ): _a : int = key.replace("""t5""" , """language""" ) _a : int = val # read in qv biases read_in_q_v_bias(UpperCamelCase__ , UpperCamelCase__ ) # note: weights get loaded in torch.float32 by default hf_model.load_state_dict(UpperCamelCase__ , strict=UpperCamelCase__ ) _a : Optional[Any] = load_demo_image() _a : int = """What is unusual about this image?""" # create processor _a : Union[str, Any] = BlipImageProcessor( size={"""height""": image_size, """width""": image_size} , image_mean=UpperCamelCase__ , image_std=UpperCamelCase__ ) _a : Optional[Any] = InstructBlipProcessor( image_processor=UpperCamelCase__ , tokenizer=UpperCamelCase__ , qformer_tokenizer=UpperCamelCase__ , ) _a : Union[str, Any] = processor(images=UpperCamelCase__ , text=UpperCamelCase__ , return_tensors="""pt""" ).to(UpperCamelCase__ ) # make sure processor creates exact same pixel values _a : str = vis_processors["""eval"""](UpperCamelCase__ ).unsqueeze(0 ).to(UpperCamelCase__ ) _a : int = inputs.pixel_values assert torch.allclose(original_pixel_values.to(pixel_values.device ) , UpperCamelCase__ ) original_model.to(UpperCamelCase__ ) hf_model.to(UpperCamelCase__ ) with torch.no_grad(): if "vicuna" in model_name: _a : List[str] = original_model({"""image""": original_pixel_values, """text_input""": [prompt]} ).logits _a : Optional[int] = hf_model(**UpperCamelCase__ ).logits else: _a : List[str] = original_model( {"""image""": original_pixel_values, """text_input""": [prompt], """text_output""": ["""\n"""]} ).logits _a : int = tokenizer("""\n""" , return_tensors="""pt""" ).input_ids.to(UpperCamelCase__ ) _a : Optional[int] = label_input_ids.masked_fill(label_input_ids == tokenizer.pad_token_id , -1_0_0 ) _a : Any = hf_model(**UpperCamelCase__ , labels=UpperCamelCase__ ).logits print("""First values of original logits:""" , original_logits[0, :3, :3] ) print("""First values of HF logits:""" , logits[0, :3, :3] ) # assert values assert original_logits.shape == logits.shape _a : Union[str, Any] = 1e-4 if """vicuna""" in model_name else 1e-5 assert torch.allclose(original_logits.to(logits.device ) , UpperCamelCase__ , atol=UpperCamelCase__ ) print("""Looks ok!""" ) print("""Generating with original model...""" ) _a : int = original_model.generate({"""image""": original_pixel_values, """prompt""": prompt} , num_beams=5 ) # important: we need to cast the weights of the HF model to the appropriate type print("""Generating with HF model...""" ) _a : Union[str, Any] = hf_model.generate( **UpperCamelCase__ , do_sample=UpperCamelCase__ , num_beams=5 , max_length=2_5_6 , min_length=1 , top_p=0.9 , repetition_penalty=1.5 , length_penalty=1.0 , temperature=1 , ) if "vicuna" in model_name: # convert output id 0 to 2 (eos_token_id) # TODO add this in the generate method? _a : List[str] = 2 print("""Original generation:""" , UpperCamelCase__ ) _a : List[Any] = processor.batch_decode(UpperCamelCase__ , skip_special_tokens=UpperCamelCase__ ) _a : Optional[Any] = [text.strip() for text in output_text] print("""HF generation:""" , UpperCamelCase__ ) if pytorch_dump_folder_path is not None: processor.save_pretrained(UpperCamelCase__ ) hf_model.save_pretrained(UpperCamelCase__ ) if push_to_hub: processor.push_to_hub(F"""Salesforce/{model_name}""" ) hf_model.push_to_hub(F"""Salesforce/{model_name}""" ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() _snake_case = [ 'instructblip-vicuna-7b', 'instructblip-vicuna-13b', 'instructblip-flan-t5-xl', 'instructblip-flan-t5-xxl', ] parser.add_argument( '--model_name', default='instructblip-flan-t5-xl', choices=choices, type=str, help='Path to hf config.json of model to convert', ) parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument( '--push_to_hub', action='store_true', help='Whether to push the model and processor to the hub after converting', ) _snake_case = parser.parse_args() convert_blipa_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
294
"""simple docstring""" def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' # Check if the input is valid if not len(UpperCamelCase__ ) == len(UpperCamelCase__ ) == 3: raise ValueError("""Please enter a valid equation.""" ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError("""Both a & b of two equations can't be zero.""" ) # Extract the coefficients _a , _a , _a : Any = equationa _a , _a , _a : Tuple = equationa # Calculate the determinants of the matrices _a : int = aa * ba - aa * ba _a : str = ca * ba - ca * ba _a : str = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError("""Infinite solutions. (Consistent system)""" ) else: raise ValueError("""No solution. (Inconsistent system)""" ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: _a : Dict = determinant_x / determinant _a : str = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
294
1